var/home/core/zuul-output/0000755000175000017500000000000015067443324014535 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067465025015503 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006017550215067465014017713 0ustar rootrootOct 02 09:36:50 crc systemd[1]: Starting Kubernetes Kubelet... Oct 02 09:36:50 crc restorecon[4676]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:36:50 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:51 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:36:52 crc restorecon[4676]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:36:52 crc restorecon[4676]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 02 09:36:53 crc kubenswrapper[4771]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 09:36:53 crc kubenswrapper[4771]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 02 09:36:53 crc kubenswrapper[4771]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 09:36:53 crc kubenswrapper[4771]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 09:36:53 crc kubenswrapper[4771]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 02 09:36:53 crc kubenswrapper[4771]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.083465 4771 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090777 4771 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090804 4771 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090813 4771 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090819 4771 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090825 4771 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090829 4771 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090833 4771 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090838 4771 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090843 4771 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090858 4771 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090862 4771 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090868 4771 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090874 4771 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090878 4771 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090882 4771 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090895 4771 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090899 4771 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090904 4771 feature_gate.go:330] unrecognized feature gate: Example Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090908 4771 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090912 4771 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090917 4771 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090921 4771 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090926 4771 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090930 4771 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090936 4771 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090942 4771 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090949 4771 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090954 4771 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090959 4771 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090963 4771 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090967 4771 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090972 4771 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090976 4771 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090980 4771 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090984 4771 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090989 4771 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090993 4771 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.090997 4771 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091004 4771 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091010 4771 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091015 4771 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091021 4771 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091026 4771 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091031 4771 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091035 4771 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091041 4771 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091044 4771 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091049 4771 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091053 4771 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091057 4771 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091061 4771 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091065 4771 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091070 4771 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091075 4771 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091080 4771 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091086 4771 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091091 4771 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091095 4771 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091100 4771 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091104 4771 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091109 4771 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091113 4771 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091120 4771 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091140 4771 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091146 4771 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091151 4771 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091155 4771 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091160 4771 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091165 4771 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091170 4771 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.091175 4771 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091309 4771 flags.go:64] FLAG: --address="0.0.0.0" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091325 4771 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091336 4771 flags.go:64] FLAG: --anonymous-auth="true" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091344 4771 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091351 4771 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091357 4771 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091364 4771 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091372 4771 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091378 4771 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091384 4771 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091390 4771 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091396 4771 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091402 4771 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091407 4771 flags.go:64] FLAG: --cgroup-root="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091413 4771 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091418 4771 flags.go:64] FLAG: --client-ca-file="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091424 4771 flags.go:64] FLAG: --cloud-config="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091430 4771 flags.go:64] FLAG: --cloud-provider="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091435 4771 flags.go:64] FLAG: --cluster-dns="[]" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091444 4771 flags.go:64] FLAG: --cluster-domain="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091449 4771 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091455 4771 flags.go:64] FLAG: --config-dir="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091460 4771 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091466 4771 flags.go:64] FLAG: --container-log-max-files="5" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091475 4771 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091481 4771 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091487 4771 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091493 4771 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091499 4771 flags.go:64] FLAG: --contention-profiling="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091504 4771 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091510 4771 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091517 4771 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091523 4771 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091530 4771 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091536 4771 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091541 4771 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091547 4771 flags.go:64] FLAG: --enable-load-reader="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091554 4771 flags.go:64] FLAG: --enable-server="true" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091559 4771 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091567 4771 flags.go:64] FLAG: --event-burst="100" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091573 4771 flags.go:64] FLAG: --event-qps="50" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091579 4771 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091586 4771 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091592 4771 flags.go:64] FLAG: --eviction-hard="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091599 4771 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091605 4771 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091611 4771 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091617 4771 flags.go:64] FLAG: --eviction-soft="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091623 4771 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091628 4771 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091634 4771 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091640 4771 flags.go:64] FLAG: --experimental-mounter-path="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091645 4771 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091651 4771 flags.go:64] FLAG: --fail-swap-on="true" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091656 4771 flags.go:64] FLAG: --feature-gates="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091664 4771 flags.go:64] FLAG: --file-check-frequency="20s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091670 4771 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091677 4771 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091683 4771 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091689 4771 flags.go:64] FLAG: --healthz-port="10248" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091695 4771 flags.go:64] FLAG: --help="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091701 4771 flags.go:64] FLAG: --hostname-override="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091706 4771 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091712 4771 flags.go:64] FLAG: --http-check-frequency="20s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091719 4771 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091724 4771 flags.go:64] FLAG: --image-credential-provider-config="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091730 4771 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091737 4771 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091742 4771 flags.go:64] FLAG: --image-service-endpoint="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091748 4771 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091753 4771 flags.go:64] FLAG: --kube-api-burst="100" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091759 4771 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091766 4771 flags.go:64] FLAG: --kube-api-qps="50" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091774 4771 flags.go:64] FLAG: --kube-reserved="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091780 4771 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091785 4771 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091791 4771 flags.go:64] FLAG: --kubelet-cgroups="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091797 4771 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091802 4771 flags.go:64] FLAG: --lock-file="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091808 4771 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091814 4771 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091819 4771 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091828 4771 flags.go:64] FLAG: --log-json-split-stream="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091834 4771 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091840 4771 flags.go:64] FLAG: --log-text-split-stream="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091846 4771 flags.go:64] FLAG: --logging-format="text" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091851 4771 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091857 4771 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091863 4771 flags.go:64] FLAG: --manifest-url="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091869 4771 flags.go:64] FLAG: --manifest-url-header="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091876 4771 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091882 4771 flags.go:64] FLAG: --max-open-files="1000000" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091889 4771 flags.go:64] FLAG: --max-pods="110" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091895 4771 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091900 4771 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091908 4771 flags.go:64] FLAG: --memory-manager-policy="None" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091914 4771 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091920 4771 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091926 4771 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091931 4771 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091946 4771 flags.go:64] FLAG: --node-status-max-images="50" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091952 4771 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091958 4771 flags.go:64] FLAG: --oom-score-adj="-999" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091964 4771 flags.go:64] FLAG: --pod-cidr="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091970 4771 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091989 4771 flags.go:64] FLAG: --pod-manifest-path="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.091995 4771 flags.go:64] FLAG: --pod-max-pids="-1" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092001 4771 flags.go:64] FLAG: --pods-per-core="0" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092007 4771 flags.go:64] FLAG: --port="10250" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092015 4771 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092020 4771 flags.go:64] FLAG: --provider-id="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092026 4771 flags.go:64] FLAG: --qos-reserved="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092031 4771 flags.go:64] FLAG: --read-only-port="10255" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092037 4771 flags.go:64] FLAG: --register-node="true" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092043 4771 flags.go:64] FLAG: --register-schedulable="true" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092049 4771 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092059 4771 flags.go:64] FLAG: --registry-burst="10" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092065 4771 flags.go:64] FLAG: --registry-qps="5" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092071 4771 flags.go:64] FLAG: --reserved-cpus="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092076 4771 flags.go:64] FLAG: --reserved-memory="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092083 4771 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092089 4771 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092095 4771 flags.go:64] FLAG: --rotate-certificates="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092101 4771 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092106 4771 flags.go:64] FLAG: --runonce="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092162 4771 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092169 4771 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092175 4771 flags.go:64] FLAG: --seccomp-default="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092181 4771 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092187 4771 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092193 4771 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092199 4771 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092204 4771 flags.go:64] FLAG: --storage-driver-password="root" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092210 4771 flags.go:64] FLAG: --storage-driver-secure="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092216 4771 flags.go:64] FLAG: --storage-driver-table="stats" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092221 4771 flags.go:64] FLAG: --storage-driver-user="root" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092226 4771 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092236 4771 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092242 4771 flags.go:64] FLAG: --system-cgroups="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092248 4771 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092258 4771 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092263 4771 flags.go:64] FLAG: --tls-cert-file="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092269 4771 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092281 4771 flags.go:64] FLAG: --tls-min-version="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092287 4771 flags.go:64] FLAG: --tls-private-key-file="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092293 4771 flags.go:64] FLAG: --topology-manager-policy="none" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092298 4771 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092304 4771 flags.go:64] FLAG: --topology-manager-scope="container" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092309 4771 flags.go:64] FLAG: --v="2" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092318 4771 flags.go:64] FLAG: --version="false" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092326 4771 flags.go:64] FLAG: --vmodule="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092333 4771 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092339 4771 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092477 4771 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092485 4771 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092493 4771 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092499 4771 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092505 4771 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092511 4771 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092523 4771 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092530 4771 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092538 4771 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092543 4771 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092548 4771 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092553 4771 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092558 4771 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092562 4771 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092567 4771 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092572 4771 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092580 4771 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092585 4771 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092590 4771 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092594 4771 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092599 4771 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092603 4771 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092607 4771 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092614 4771 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092618 4771 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092623 4771 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092628 4771 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092632 4771 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092638 4771 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092642 4771 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092647 4771 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092651 4771 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092656 4771 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092660 4771 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092665 4771 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092669 4771 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092674 4771 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092678 4771 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092686 4771 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092691 4771 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092696 4771 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092702 4771 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092706 4771 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092711 4771 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092715 4771 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092719 4771 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092724 4771 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092728 4771 feature_gate.go:330] unrecognized feature gate: Example Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092737 4771 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092742 4771 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092746 4771 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092751 4771 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092756 4771 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092760 4771 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092765 4771 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092769 4771 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092774 4771 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092778 4771 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092783 4771 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092788 4771 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092792 4771 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092797 4771 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092801 4771 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092806 4771 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092825 4771 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092831 4771 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092835 4771 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092840 4771 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092845 4771 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092851 4771 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.092860 4771 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.092876 4771 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.105716 4771 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.105788 4771 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105885 4771 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105898 4771 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105921 4771 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105926 4771 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105931 4771 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105935 4771 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105939 4771 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105943 4771 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105947 4771 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105950 4771 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105954 4771 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105958 4771 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105962 4771 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105966 4771 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105974 4771 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105981 4771 feature_gate.go:330] unrecognized feature gate: Example Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105985 4771 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105990 4771 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105995 4771 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.105999 4771 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106003 4771 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106008 4771 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106012 4771 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106016 4771 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106020 4771 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106024 4771 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106029 4771 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106033 4771 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106038 4771 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106042 4771 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106046 4771 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106050 4771 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106058 4771 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106067 4771 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106088 4771 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106107 4771 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106112 4771 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106117 4771 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106122 4771 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106139 4771 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106144 4771 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106148 4771 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106152 4771 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106156 4771 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106162 4771 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106166 4771 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106171 4771 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106177 4771 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106182 4771 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106187 4771 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106191 4771 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106196 4771 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106201 4771 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106206 4771 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106210 4771 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106214 4771 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106218 4771 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106222 4771 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106225 4771 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106229 4771 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106232 4771 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106237 4771 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106243 4771 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106248 4771 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106252 4771 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106258 4771 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106263 4771 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106268 4771 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106272 4771 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106277 4771 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106290 4771 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.106298 4771 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106479 4771 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106494 4771 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106500 4771 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106506 4771 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106512 4771 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106518 4771 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106523 4771 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106529 4771 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106534 4771 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106545 4771 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106553 4771 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106558 4771 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106563 4771 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106568 4771 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106572 4771 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106577 4771 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106582 4771 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106587 4771 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106592 4771 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106596 4771 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106601 4771 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106605 4771 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106613 4771 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106622 4771 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106627 4771 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106633 4771 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106639 4771 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106645 4771 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106651 4771 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106656 4771 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106661 4771 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106666 4771 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106671 4771 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106676 4771 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106690 4771 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106697 4771 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106704 4771 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106709 4771 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106715 4771 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106720 4771 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106725 4771 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106730 4771 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106735 4771 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106740 4771 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106745 4771 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106749 4771 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106754 4771 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106759 4771 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106764 4771 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106769 4771 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106774 4771 feature_gate.go:330] unrecognized feature gate: Example Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106778 4771 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106783 4771 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106788 4771 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106793 4771 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106799 4771 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106804 4771 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106810 4771 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106816 4771 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106823 4771 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106828 4771 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106834 4771 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106840 4771 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106844 4771 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106849 4771 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106854 4771 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106859 4771 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106865 4771 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106870 4771 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106875 4771 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.106888 4771 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.106897 4771 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.108297 4771 server.go:940] "Client rotation is on, will bootstrap in background" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.113947 4771 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.114101 4771 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.116452 4771 server.go:997] "Starting client certificate rotation" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.116527 4771 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.116873 4771 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-07 12:54:46.888011818 +0000 UTC Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.117030 4771 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1587h17m53.77098709s for next certificate rotation Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.171371 4771 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.173302 4771 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.257056 4771 log.go:25] "Validated CRI v1 runtime API" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.366905 4771 log.go:25] "Validated CRI v1 image API" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.369197 4771 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.383497 4771 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-02-09-32-04-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.383543 4771 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.416189 4771 manager.go:217] Machine: {Timestamp:2025-10-02 09:36:53.407188142 +0000 UTC m=+1.054873289 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:d79e76cf-cd2e-4589-9fea-018e1e59062c BootID:bea03645-5dbd-471c-a261-bbba6e916461 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:8a:77:d0 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:8a:77:d0 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:2f:7f:e8 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:71:4c:a2 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:af:41:cc Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e4:36:7c Speed:-1 Mtu:1496} {Name:eth10 MacAddress:aa:ae:0f:d3:78:63 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:66:7d:20:ec:16:c1 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.416597 4771 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.416856 4771 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.420829 4771 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.421218 4771 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.421278 4771 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.421605 4771 topology_manager.go:138] "Creating topology manager with none policy" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.421623 4771 container_manager_linux.go:303] "Creating device plugin manager" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.422447 4771 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.422509 4771 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.422779 4771 state_mem.go:36] "Initialized new in-memory state store" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.422928 4771 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.462848 4771 kubelet.go:418] "Attempting to sync node with API server" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.462888 4771 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.462924 4771 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.462946 4771 kubelet.go:324] "Adding apiserver pod source" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.462966 4771 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.468024 4771 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.470858 4771 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.472192 4771 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.472321 4771 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.472187 4771 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.472414 4771 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.477725 4771 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.480219 4771 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.480263 4771 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.480278 4771 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.480297 4771 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.480329 4771 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.480342 4771 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.480355 4771 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.480376 4771 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.480391 4771 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.480404 4771 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.480447 4771 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.480460 4771 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.486464 4771 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.487688 4771 server.go:1280] "Started kubelet" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.487956 4771 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:53 crc systemd[1]: Started Kubernetes Kubelet. Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.493294 4771 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.493300 4771 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.494213 4771 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.497568 4771 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.497610 4771 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.497848 4771 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.497977 4771 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 18:16:23.572378727 +0000 UTC Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.498077 4771 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1976h39m30.074305663s for next certificate rotation Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.498042 4771 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.498113 4771 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.498046 4771 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.524185 4771 factory.go:55] Registering systemd factory Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.524513 4771 factory.go:221] Registration of the systemd container factory successfully Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.525282 4771 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.525423 4771 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.525561 4771 factory.go:153] Registering CRI-O factory Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.525634 4771 factory.go:221] Registration of the crio container factory successfully Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.526981 4771 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.527010 4771 factory.go:103] Registering Raw factory Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.527029 4771 manager.go:1196] Started watching for new ooms in manager Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.527916 4771 manager.go:319] Starting recovery of all containers Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.529267 4771 server.go:460] "Adding debug handlers to kubelet server" Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.529446 4771 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.53:6443: connect: connection refused" interval="200ms" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581053 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581167 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581198 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581213 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581271 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581284 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581302 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581320 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581357 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581381 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581392 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581419 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581430 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581455 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581474 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581493 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581517 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581531 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581565 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581579 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581600 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581627 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581648 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581877 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581907 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581928 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.581963 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.582032 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.582050 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.582078 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.582095 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.582118 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.583268 4771 manager.go:324] Recovery completed Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.595863 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.598082 4771 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.598369 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.598506 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.598557 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.599688 4771 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.599705 4771 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.599728 4771 state_mem.go:36] "Initialized new in-memory state store" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.675763 4771 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.679703 4771 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.679757 4771 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.679913 4771 kubelet.go:2335] "Starting kubelet main sync loop" Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.680513 4771 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 02 09:36:53 crc kubenswrapper[4771]: W1002 09:36:53.681536 4771 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.681603 4771 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.582826 4771 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.53:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186aa2fd5b877d4e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 09:36:53.487631694 +0000 UTC m=+1.135316851,LastTimestamp:2025-10-02 09:36:53.487631694 +0000 UTC m=+1.135316851,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685633 4771 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685719 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685746 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685767 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685785 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685803 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685825 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685843 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685859 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685874 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685890 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685907 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685921 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685936 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685952 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685969 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.685987 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686002 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686018 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686034 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686054 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686081 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686123 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686170 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686188 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686203 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686220 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686235 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686250 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686267 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686285 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686299 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686329 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686344 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686360 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686374 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686390 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686404 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686420 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686434 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686452 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686466 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686491 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686508 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686524 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686540 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686556 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686572 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686590 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686602 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686614 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686635 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686648 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686658 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686670 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686681 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686691 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686701 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686713 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686723 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686733 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686746 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686757 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686770 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686779 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686789 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686800 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686811 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686822 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686832 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686843 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686853 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686863 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686880 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686892 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686915 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686939 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686955 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686971 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.686987 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687001 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687014 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687047 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687062 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687077 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687088 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687099 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687111 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687122 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687182 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687194 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687206 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687218 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687229 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687240 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687253 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687264 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687276 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687287 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687301 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687313 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687324 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687335 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687346 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687356 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687366 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687376 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687387 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687398 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687408 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687419 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687431 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687450 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687472 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687486 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687500 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687510 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687521 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687531 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687543 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687554 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687566 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687578 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687591 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687603 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687614 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687624 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687636 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687652 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687667 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687681 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687691 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687702 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687712 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687723 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687735 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687747 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687758 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687770 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687787 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687798 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687809 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687819 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687833 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687851 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687868 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687881 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687900 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687914 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687928 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687942 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687958 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687970 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687985 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.687996 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688011 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688033 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688049 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688064 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688079 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688092 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688106 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688122 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688160 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688177 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688196 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688209 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688219 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688231 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688241 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688252 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688265 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688277 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688289 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688299 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688311 4771 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688321 4771 reconstruct.go:97] "Volume reconstruction finished" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.688331 4771 reconciler.go:26] "Reconciler: start to sync state" Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.698314 4771 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.731055 4771 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.53:6443: connect: connection refused" interval="400ms" Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.781072 4771 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.798644 4771 policy_none.go:49] "None policy: Start" Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.799004 4771 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.800523 4771 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.800553 4771 state_mem.go:35] "Initializing new in-memory state store" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.875612 4771 manager.go:334] "Starting Device Plugin manager" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.876437 4771 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.876455 4771 server.go:79] "Starting device plugin registration server" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.876982 4771 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.876998 4771 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.877269 4771 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.877351 4771 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.877359 4771 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.884666 4771 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.978035 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.979329 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.979359 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.979371 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.979400 4771 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:36:53 crc kubenswrapper[4771]: E1002 09:36:53.979865 4771 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.53:6443: connect: connection refused" node="crc" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.982231 4771 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.982435 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.983883 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.983928 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.983941 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.984185 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.984368 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.984412 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.985137 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.985162 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.985172 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.985174 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.985193 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.985204 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.985315 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.985416 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.985442 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.986060 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.986134 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.986151 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.986314 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.986336 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.986348 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.986352 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.986702 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.986813 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.987567 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.987600 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.987613 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.987721 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.987891 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.987928 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.988403 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.988425 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.988434 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.988959 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.988992 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.989014 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.989024 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.988992 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.989173 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.989477 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.989512 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.990542 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.990573 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:53 crc kubenswrapper[4771]: I1002 09:36:53.990599 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093075 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093441 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093517 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093570 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093646 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093691 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093711 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093735 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093749 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093765 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093782 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093801 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093824 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093844 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.093863 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: E1002 09:36:54.132525 4771 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.53:6443: connect: connection refused" interval="800ms" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.180307 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.182255 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.182329 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.182385 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.182428 4771 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:36:54 crc kubenswrapper[4771]: E1002 09:36:54.183218 4771 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.53:6443: connect: connection refused" node="crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195299 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195358 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195391 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195440 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195469 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195493 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195495 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195565 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195508 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195520 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195613 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195624 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195634 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195578 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195733 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195765 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195785 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195806 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195830 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195834 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195853 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195895 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195913 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.195967 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.196083 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.196164 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.196175 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.196190 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.196230 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.196290 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.317547 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.338769 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.357833 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.401062 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.401986 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:36:54 crc kubenswrapper[4771]: W1002 09:36:54.456955 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-4d75a57eea8a8e9678fa4e3e83d15d705d02cfcf292354cbd0d8e3750a2e0dbd WatchSource:0}: Error finding container 4d75a57eea8a8e9678fa4e3e83d15d705d02cfcf292354cbd0d8e3750a2e0dbd: Status 404 returned error can't find the container with id 4d75a57eea8a8e9678fa4e3e83d15d705d02cfcf292354cbd0d8e3750a2e0dbd Oct 02 09:36:54 crc kubenswrapper[4771]: W1002 09:36:54.461963 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-f130203e3672a071cac63eec38fff0ef0d68e981aa602bc6905acc6656b7b063 WatchSource:0}: Error finding container f130203e3672a071cac63eec38fff0ef0d68e981aa602bc6905acc6656b7b063: Status 404 returned error can't find the container with id f130203e3672a071cac63eec38fff0ef0d68e981aa602bc6905acc6656b7b063 Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.489482 4771 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:54 crc kubenswrapper[4771]: W1002 09:36:54.492407 4771 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:54 crc kubenswrapper[4771]: E1002 09:36:54.492514 4771 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.583935 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.586335 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.586395 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.586407 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.586443 4771 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:36:54 crc kubenswrapper[4771]: E1002 09:36:54.587120 4771 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.53:6443: connect: connection refused" node="crc" Oct 02 09:36:54 crc kubenswrapper[4771]: W1002 09:36:54.659046 4771 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:54 crc kubenswrapper[4771]: E1002 09:36:54.659171 4771 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.686761 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4d75a57eea8a8e9678fa4e3e83d15d705d02cfcf292354cbd0d8e3750a2e0dbd"} Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.688487 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"31ffb5669cb3d52e596e90b8c4da82bd43f1754ca8644e8f9049b5a6e2a8eaa7"} Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.689615 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"846a1e21041d007947aea6d260d3434893ef28aab125c94f1506208c99580380"} Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.690671 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9f9c5d597bde8a5983600c780884e28baf9a4521a34b88f8d3ee378bfb60d36a"} Oct 02 09:36:54 crc kubenswrapper[4771]: I1002 09:36:54.691881 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f130203e3672a071cac63eec38fff0ef0d68e981aa602bc6905acc6656b7b063"} Oct 02 09:36:54 crc kubenswrapper[4771]: W1002 09:36:54.766821 4771 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:54 crc kubenswrapper[4771]: E1002 09:36:54.766931 4771 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:36:54 crc kubenswrapper[4771]: W1002 09:36:54.890169 4771 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:54 crc kubenswrapper[4771]: E1002 09:36:54.890256 4771 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:36:54 crc kubenswrapper[4771]: E1002 09:36:54.934238 4771 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.53:6443: connect: connection refused" interval="1.6s" Oct 02 09:36:55 crc kubenswrapper[4771]: I1002 09:36:55.387526 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:55 crc kubenswrapper[4771]: I1002 09:36:55.389096 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:55 crc kubenswrapper[4771]: I1002 09:36:55.389164 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:55 crc kubenswrapper[4771]: I1002 09:36:55.389177 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:55 crc kubenswrapper[4771]: I1002 09:36:55.389212 4771 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:36:55 crc kubenswrapper[4771]: E1002 09:36:55.389744 4771 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.53:6443: connect: connection refused" node="crc" Oct 02 09:36:55 crc kubenswrapper[4771]: I1002 09:36:55.489402 4771 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:56 crc kubenswrapper[4771]: W1002 09:36:56.226083 4771 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:56 crc kubenswrapper[4771]: E1002 09:36:56.226180 4771 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.489504 4771 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:56 crc kubenswrapper[4771]: E1002 09:36:56.535240 4771 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.53:6443: connect: connection refused" interval="3.2s" Oct 02 09:36:56 crc kubenswrapper[4771]: W1002 09:36:56.629500 4771 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:56 crc kubenswrapper[4771]: E1002 09:36:56.629569 4771 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.698415 4771 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="5f4e8484386c5d2b52033e90f3e92d987f9e18ba5ea7e0605ddbfc96215ff3c1" exitCode=0 Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.698497 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"5f4e8484386c5d2b52033e90f3e92d987f9e18ba5ea7e0605ddbfc96215ff3c1"} Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.698586 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.702925 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.702975 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.702995 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.704218 4771 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa" exitCode=0 Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.704303 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa"} Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.704353 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.705404 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.705483 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.705556 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.706856 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510"} Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.709905 4771 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7" exitCode=0 Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.709985 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7"} Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.709999 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.711523 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.711566 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.711577 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.713011 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.713423 4771 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d" exitCode=0 Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.713495 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d"} Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.713680 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.714255 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.714291 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.714306 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.714911 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.714938 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.714951 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:56 crc kubenswrapper[4771]: W1002 09:36:56.730025 4771 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:56 crc kubenswrapper[4771]: E1002 09:36:56.730168 4771 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.990289 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.994622 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.994672 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.994682 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:56 crc kubenswrapper[4771]: I1002 09:36:56.994714 4771 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:36:56 crc kubenswrapper[4771]: E1002 09:36:56.995370 4771 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.53:6443: connect: connection refused" node="crc" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.488942 4771 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.721985 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"92ac7ad8673c663a0665b8373b9b53d2d223d1ce7e74e04783b1d63c2f01a83f"} Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.722152 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.723331 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.723378 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.723400 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.727813 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"606cc0b4bdb6eda2e3d722157debd219232d6c3c1877e4c10affe91d0bbb12a0"} Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.727864 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"82587f84343f1807330d05aa3632a5acdb5462cc6f11ef6cd678d32164231284"} Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.727877 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6047ebfe8e452e54f1e3fde9af41485e5a527210666b8b94c395c8ceded02c42"} Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.727891 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.733596 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.733645 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.733656 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.735606 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92"} Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.735664 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785"} Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.735675 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93"} Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.735772 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.737437 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.737489 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.737498 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.741326 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb"} Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.741393 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a"} Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.741411 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a"} Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.741424 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25"} Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.743015 4771 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f" exitCode=0 Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.743098 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f"} Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.743277 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.744159 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.744190 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:57 crc kubenswrapper[4771]: I1002 09:36:57.744201 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:57 crc kubenswrapper[4771]: W1002 09:36:57.919620 4771 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:57 crc kubenswrapper[4771]: E1002 09:36:57.919712 4771 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.489899 4771 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.749245 4771 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522" exitCode=0 Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.749375 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522"} Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.749419 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.750403 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.750439 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.750453 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.753091 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3e691876c1e3985b10cd66b3c091e672b16b3ea560de8ca7c361c9c967fb1518"} Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.753191 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.753240 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.753266 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.753200 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.753240 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.754212 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.754229 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.754244 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.754250 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.754262 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.754305 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.754389 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.754413 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.754425 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.758524 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.758589 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:58 crc kubenswrapper[4771]: I1002 09:36:58.758609 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.618857 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.650040 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.757550 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.759387 4771 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3e691876c1e3985b10cd66b3c091e672b16b3ea560de8ca7c361c9c967fb1518" exitCode=255 Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.759472 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3e691876c1e3985b10cd66b3c091e672b16b3ea560de8ca7c361c9c967fb1518"} Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.759610 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.760732 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.760814 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.760839 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.761934 4771 scope.go:117] "RemoveContainer" containerID="3e691876c1e3985b10cd66b3c091e672b16b3ea560de8ca7c361c9c967fb1518" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.764372 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470"} Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.764430 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.764436 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540"} Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.764457 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7"} Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.764484 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.764430 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.765334 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.765359 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.765372 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.765394 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.765462 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:36:59 crc kubenswrapper[4771]: I1002 09:36:59.765482 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.160768 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.196223 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.197616 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.197675 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.197688 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.197719 4771 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.638807 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.769412 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.772075 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355"} Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.772180 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.772220 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.773105 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.773148 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.773160 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.776207 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904"} Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.776314 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.777184 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.777219 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:00 crc kubenswrapper[4771]: I1002 09:37:00.777227 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:01 crc kubenswrapper[4771]: I1002 09:37:01.784240 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:01 crc kubenswrapper[4771]: I1002 09:37:01.784790 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:01 crc kubenswrapper[4771]: I1002 09:37:01.784914 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e"} Oct 02 09:37:01 crc kubenswrapper[4771]: I1002 09:37:01.785041 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:37:01 crc kubenswrapper[4771]: I1002 09:37:01.785664 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:01 crc kubenswrapper[4771]: I1002 09:37:01.785695 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:01 crc kubenswrapper[4771]: I1002 09:37:01.785704 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:01 crc kubenswrapper[4771]: I1002 09:37:01.785953 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:01 crc kubenswrapper[4771]: I1002 09:37:01.786030 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:01 crc kubenswrapper[4771]: I1002 09:37:01.786086 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:02 crc kubenswrapper[4771]: I1002 09:37:02.787391 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:02 crc kubenswrapper[4771]: I1002 09:37:02.787513 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:02 crc kubenswrapper[4771]: I1002 09:37:02.789925 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:02 crc kubenswrapper[4771]: I1002 09:37:02.789972 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:02 crc kubenswrapper[4771]: I1002 09:37:02.789983 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:02 crc kubenswrapper[4771]: I1002 09:37:02.792981 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:02 crc kubenswrapper[4771]: I1002 09:37:02.793049 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:02 crc kubenswrapper[4771]: I1002 09:37:02.793067 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:03 crc kubenswrapper[4771]: I1002 09:37:03.538940 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:37:03 crc kubenswrapper[4771]: I1002 09:37:03.539247 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:03 crc kubenswrapper[4771]: I1002 09:37:03.540521 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:03 crc kubenswrapper[4771]: I1002 09:37:03.540570 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:03 crc kubenswrapper[4771]: I1002 09:37:03.540584 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:03 crc kubenswrapper[4771]: E1002 09:37:03.884793 4771 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 09:37:04 crc kubenswrapper[4771]: I1002 09:37:04.529781 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:37:04 crc kubenswrapper[4771]: I1002 09:37:04.529985 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:04 crc kubenswrapper[4771]: I1002 09:37:04.531706 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:04 crc kubenswrapper[4771]: I1002 09:37:04.531767 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:04 crc kubenswrapper[4771]: I1002 09:37:04.531781 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:06 crc kubenswrapper[4771]: I1002 09:37:06.539179 4771 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 09:37:06 crc kubenswrapper[4771]: I1002 09:37:06.539312 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 09:37:06 crc kubenswrapper[4771]: I1002 09:37:06.634939 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 02 09:37:06 crc kubenswrapper[4771]: I1002 09:37:06.635660 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:06 crc kubenswrapper[4771]: I1002 09:37:06.637729 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:06 crc kubenswrapper[4771]: I1002 09:37:06.637781 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:06 crc kubenswrapper[4771]: I1002 09:37:06.637796 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:06 crc kubenswrapper[4771]: I1002 09:37:06.802218 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 02 09:37:06 crc kubenswrapper[4771]: I1002 09:37:06.802511 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:06 crc kubenswrapper[4771]: I1002 09:37:06.803869 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:06 crc kubenswrapper[4771]: I1002 09:37:06.803903 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:06 crc kubenswrapper[4771]: I1002 09:37:06.803918 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:09 crc kubenswrapper[4771]: I1002 09:37:09.174248 4771 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 09:37:09 crc kubenswrapper[4771]: I1002 09:37:09.174366 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 09:37:09 crc kubenswrapper[4771]: I1002 09:37:09.181199 4771 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 09:37:09 crc kubenswrapper[4771]: I1002 09:37:09.181264 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.644395 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.644588 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.644969 4771 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.645015 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.645907 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.645958 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.645971 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.648583 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.808883 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.809850 4771 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.809954 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.814092 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.814206 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:10 crc kubenswrapper[4771]: I1002 09:37:10.814226 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:13 crc kubenswrapper[4771]: E1002 09:37:13.885036 4771 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.158862 4771 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.160548 4771 trace.go:236] Trace[843921158]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 09:37:02.390) (total time: 11770ms): Oct 02 09:37:14 crc kubenswrapper[4771]: Trace[843921158]: ---"Objects listed" error: 11770ms (09:37:14.160) Oct 02 09:37:14 crc kubenswrapper[4771]: Trace[843921158]: [11.770207458s] [11.770207458s] END Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.160723 4771 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.164828 4771 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.181997 4771 trace.go:236] Trace[490466299]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 09:37:02.956) (total time: 11225ms): Oct 02 09:37:14 crc kubenswrapper[4771]: Trace[490466299]: ---"Objects listed" error: 11225ms (09:37:14.181) Oct 02 09:37:14 crc kubenswrapper[4771]: Trace[490466299]: [11.225281903s] [11.225281903s] END Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.182044 4771 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.182308 4771 trace.go:236] Trace[1881779095]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 09:37:00.306) (total time: 13876ms): Oct 02 09:37:14 crc kubenswrapper[4771]: Trace[1881779095]: ---"Objects listed" error: 13876ms (09:37:14.182) Oct 02 09:37:14 crc kubenswrapper[4771]: Trace[1881779095]: [13.876129936s] [13.876129936s] END Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.182379 4771 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.185021 4771 trace.go:236] Trace[66837045]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 09:37:01.987) (total time: 12197ms): Oct 02 09:37:14 crc kubenswrapper[4771]: Trace[66837045]: ---"Objects listed" error: 12197ms (09:37:14.184) Oct 02 09:37:14 crc kubenswrapper[4771]: Trace[66837045]: [12.197166218s] [12.197166218s] END Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.185061 4771 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.185822 4771 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.475189 4771 apiserver.go:52] "Watching apiserver" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.479807 4771 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.480340 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.480859 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.480995 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.481044 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.481105 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.481116 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.481147 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.481045 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.481111 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.481496 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.482900 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.483270 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.483398 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.483470 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.485389 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.485462 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.485673 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.485679 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.486344 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.500118 4771 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.509194 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.524998 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.546971 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.559459 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.569251 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.579771 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.588105 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.588445 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.588528 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.588606 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.588675 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.588752 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.588811 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.588889 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.588963 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.589026 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.589181 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.589316 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.589427 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.589513 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.589586 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.589664 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.589808 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.589939 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.590071 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.590198 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.590271 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.589353 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.589702 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.589727 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.589754 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.590051 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.590344 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.590582 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.590655 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.590729 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.590799 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.590869 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.590938 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591007 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591112 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591215 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591289 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591360 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591460 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591586 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591670 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591751 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591828 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591897 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591968 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592183 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592225 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592251 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592274 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592296 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592333 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592359 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592384 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592410 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592434 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592459 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592483 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592500 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592515 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592536 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592558 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592583 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592605 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592627 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592645 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592664 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592683 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592704 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592720 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592737 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592754 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592777 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592795 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592815 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592832 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592850 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592867 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592885 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592901 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592922 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592952 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592978 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593007 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593033 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593058 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593084 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593108 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593157 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593182 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593206 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593228 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593252 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593280 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593305 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593328 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593352 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593370 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593389 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593405 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593423 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593442 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593461 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593486 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593518 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593542 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593569 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593593 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593622 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593640 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593659 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593677 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593693 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593709 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593731 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593748 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593766 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593783 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593800 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593818 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593839 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593862 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593880 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593900 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593917 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593945 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593972 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593989 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594006 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594023 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594041 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594061 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594080 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594101 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594143 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594167 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594195 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594214 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594231 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594253 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594274 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594293 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594312 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594331 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594352 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594372 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594397 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594419 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594438 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594457 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.590754 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.590837 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591062 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594476 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591314 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591186 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594556 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591400 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591490 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594572 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591818 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591945 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.591995 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.592913 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593037 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593573 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593882 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.593924 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594229 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594283 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594667 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594394 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594859 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594494 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.595933 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.596468 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.596440 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.597218 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.597267 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.597273 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.597425 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.597539 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.597630 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.597879 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.597910 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.598056 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.594584 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.598622 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.598781 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.598933 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.599012 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.599316 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.599154 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.599370 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.599500 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.599743 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.599745 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.599984 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.600039 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.600047 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.600302 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.600308 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.600395 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.600415 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.600311 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.600092 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.600618 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.600641 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.600649 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.600949 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.600999 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601036 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601069 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601101 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601145 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601171 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601097 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601204 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601235 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601263 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601291 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601318 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601344 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601374 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601403 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601408 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601436 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601465 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601493 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601519 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601532 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601523 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601597 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601628 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601659 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601796 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601895 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601919 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601941 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601963 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601981 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.601999 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602019 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602037 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602157 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602180 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602201 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602222 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602241 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602243 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602257 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602274 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602434 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602450 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602503 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602658 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602697 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602729 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602754 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602781 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602814 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602843 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602868 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602898 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602944 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602942 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.602970 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603018 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603051 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603086 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603116 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603190 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603208 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603234 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603265 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603304 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603335 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603369 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603373 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603398 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603429 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603410 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603458 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603534 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603611 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603632 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603630 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603671 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603650 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603793 4771 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603817 4771 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603839 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603861 4771 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603879 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603896 4771 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603901 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603906 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603933 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603913 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603952 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604001 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.603978 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604063 4771 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604082 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604098 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604113 4771 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604144 4771 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604157 4771 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604169 4771 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604179 4771 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604191 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604204 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604214 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604221 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604308 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604384 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604430 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604544 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.604708 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.605363 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.605481 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.605710 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.605742 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.606013 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.606048 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.606296 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.606315 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.607050 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.607084 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.607120 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.607140 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.607652 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.609443 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.609659 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.609638 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.609807 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.610172 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.610338 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.610605 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.611061 4771 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.611363 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.611675 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.611819 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.612078 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.612713 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.612788 4771 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.612880 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:15.112851056 +0000 UTC m=+22.760536113 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.612930 4771 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.613251 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:15.113242647 +0000 UTC m=+22.760927714 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.613280 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.613642 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.614323 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.614343 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.614698 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.614730 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.615699 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.616853 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.617663 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.617105 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.617740 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:37:15.117718531 +0000 UTC m=+22.765403598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.617812 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618039 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618057 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618073 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618113 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618332 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618374 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618905 4771 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618908 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618928 4771 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618945 4771 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618966 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618982 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618998 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.618989 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619022 4771 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619067 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619081 4771 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619093 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619107 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619121 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619154 4771 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619169 4771 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619182 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619195 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619209 4771 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619224 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619239 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619254 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619269 4771 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619282 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619297 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619311 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619327 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619344 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619357 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619371 4771 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619382 4771 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619393 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619405 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619416 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619427 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619438 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619450 4771 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619464 4771 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619476 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619506 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619520 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619532 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619548 4771 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619560 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619575 4771 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619587 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619601 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619586 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619877 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.619915 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.620473 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.622029 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.622056 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.622073 4771 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.622168 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:15.122148314 +0000 UTC m=+22.769833391 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.626307 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.628499 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.629237 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.630112 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.631615 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.631731 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.631942 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.632199 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.632007 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.632591 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.632457 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.632686 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.633044 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.633549 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.633557 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.633913 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.634161 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.634251 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.634390 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.634474 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.634685 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.636337 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.636371 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.638737 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.638801 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.638984 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.639048 4771 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.639186 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:15.139164366 +0000 UTC m=+22.786849603 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.639921 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.640339 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.640438 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.640459 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.641012 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.642026 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.642163 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.642624 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.643781 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.644029 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.644204 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.645563 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.645800 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.645800 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.645936 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.646209 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.646232 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.646462 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.646487 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.646601 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.646752 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.647077 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.647153 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.647521 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.647598 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.647714 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.647925 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.648007 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.648006 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.648114 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.648172 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.648196 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.648365 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.648401 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.649002 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.649201 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.650336 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.650516 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.650669 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.651096 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.651565 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.651599 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.651596 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.651667 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.652940 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.653592 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.667278 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.669651 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.678414 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.683941 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.720958 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721070 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721085 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721279 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721294 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721304 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721316 4771 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721325 4771 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721333 4771 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721342 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721352 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721362 4771 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721371 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721381 4771 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721430 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721443 4771 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721454 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721464 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721474 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721483 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721486 4771 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721532 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721545 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721558 4771 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721568 4771 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721578 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721588 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721599 4771 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721609 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721620 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721629 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721638 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721646 4771 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721657 4771 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721668 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721676 4771 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721683 4771 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721692 4771 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721701 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721710 4771 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721718 4771 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721726 4771 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721736 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721745 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721754 4771 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721763 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721771 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721780 4771 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721789 4771 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721797 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721806 4771 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721815 4771 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721824 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721832 4771 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721840 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721849 4771 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721857 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721866 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721875 4771 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721884 4771 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721892 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721900 4771 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721909 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721917 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721944 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721956 4771 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721964 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721974 4771 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721985 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.721997 4771 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722008 4771 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722016 4771 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722025 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722034 4771 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722043 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722051 4771 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722062 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722070 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722079 4771 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722089 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722098 4771 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722106 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722115 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722143 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722155 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722166 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722174 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722184 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722193 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722202 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722213 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722222 4771 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722230 4771 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722239 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722248 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722258 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722265 4771 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722306 4771 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722315 4771 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722326 4771 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722337 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722348 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722357 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722365 4771 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722374 4771 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722382 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722391 4771 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722400 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722408 4771 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722417 4771 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722426 4771 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722434 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722443 4771 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722452 4771 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722461 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722469 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722478 4771 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722487 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722496 4771 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722506 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722516 4771 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722525 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722533 4771 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722542 4771 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722550 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722558 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.722567 4771 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.801636 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.813051 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.820260 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.823813 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.824521 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.826736 4771 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355" exitCode=255 Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.826819 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355"} Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.826885 4771 scope.go:117] "RemoveContainer" containerID="3e691876c1e3985b10cd66b3c091e672b16b3ea560de8ca7c361c9c967fb1518" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.828275 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"0108ebf83fe5a15a67f221c63136e0a3e35e6abfa2312bfb36247891918525d8"} Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.839307 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.843278 4771 scope.go:117] "RemoveContainer" containerID="0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355" Oct 02 09:37:14 crc kubenswrapper[4771]: E1002 09:37:14.843594 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.846615 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 02 09:37:14 crc kubenswrapper[4771]: W1002 09:37:14.847713 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-7d4c65ae771e244a871c74cb78a03717d6238787c74e65cabbf592415c3da222 WatchSource:0}: Error finding container 7d4c65ae771e244a871c74cb78a03717d6238787c74e65cabbf592415c3da222: Status 404 returned error can't find the container with id 7d4c65ae771e244a871c74cb78a03717d6238787c74e65cabbf592415c3da222 Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.852693 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.863726 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.874472 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.887905 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.899644 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.958051 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.971737 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.974432 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.983836 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:14 crc kubenswrapper[4771]: I1002 09:37:14.995831 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.008853 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e691876c1e3985b10cd66b3c091e672b16b3ea560de8ca7c361c9c967fb1518\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:36:59Z\\\",\\\"message\\\":\\\"W1002 09:36:58.126490 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 09:36:58.126997 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759397818 cert, and key in /tmp/serving-cert-3724887552/serving-signer.crt, /tmp/serving-cert-3724887552/serving-signer.key\\\\nI1002 09:36:58.564934 1 observer_polling.go:159] Starting file observer\\\\nW1002 09:36:58.570394 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 09:36:58.570605 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:36:58.571437 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3724887552/tls.crt::/tmp/serving-cert-3724887552/tls.key\\\\\\\"\\\\nF1002 09:36:59.020480 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.020963 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.032989 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.045598 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.126381 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.126502 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.126533 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.126605 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:37:16.126577214 +0000 UTC m=+23.774262281 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.126636 4771 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.126692 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:16.126681437 +0000 UTC m=+23.774366504 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.126707 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.126699 4771 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.126835 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.126846 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:16.126818821 +0000 UTC m=+23.774503888 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.126854 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.126869 4771 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.126903 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:16.126894023 +0000 UTC m=+23.774579300 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.193033 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.204355 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.214412 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.222979 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.228006 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.228282 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.228326 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.228341 4771 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.228423 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:16.228399851 +0000 UTC m=+23.876084918 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.237733 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e691876c1e3985b10cd66b3c091e672b16b3ea560de8ca7c361c9c967fb1518\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:36:59Z\\\",\\\"message\\\":\\\"W1002 09:36:58.126490 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 09:36:58.126997 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759397818 cert, and key in /tmp/serving-cert-3724887552/serving-signer.crt, /tmp/serving-cert-3724887552/serving-signer.key\\\\nI1002 09:36:58.564934 1 observer_polling.go:159] Starting file observer\\\\nW1002 09:36:58.570394 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 09:36:58.570605 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:36:58.571437 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3724887552/tls.crt::/tmp/serving-cert-3724887552/tls.key\\\\\\\"\\\\nF1002 09:36:59.020480 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.251355 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.264359 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.274937 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.286183 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.685513 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.686183 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.687221 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.688052 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.688762 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.689357 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.690025 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.692532 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.693607 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.695456 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.696065 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.697307 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.697906 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.698667 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.699819 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.700631 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.701998 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.702506 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.703093 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.704149 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.704790 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.706063 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.706657 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.708361 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.708844 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.709534 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.710808 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.711358 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.712342 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.712852 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.714004 4771 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.714169 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.716536 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.717575 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.718033 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.720089 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.721121 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.722100 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.722788 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.723849 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.724648 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.725927 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.726714 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.727880 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.728439 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.729350 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.729850 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.731000 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.731487 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.732500 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.749721 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.750609 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.751390 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.752531 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.833433 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.835890 4771 scope.go:117] "RemoveContainer" containerID="0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355" Oct 02 09:37:15 crc kubenswrapper[4771]: E1002 09:37:15.836078 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.836450 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6"} Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.836506 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7d4c65ae771e244a871c74cb78a03717d6238787c74e65cabbf592415c3da222"} Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.838972 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa"} Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.839033 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75"} Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.839051 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ccbe7f9fce7bb46310218d7a14658fae2f910beb5d26d4b3d5294b93f5baf8b9"} Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.842483 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.851955 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.864988 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.879657 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.895114 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.914891 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.931688 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.943472 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.955020 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.970291 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:15 crc kubenswrapper[4771]: I1002 09:37:15.985257 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.000985 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.017177 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.045593 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.060687 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.076359 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.090440 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.136060 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.136227 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.136264 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.136291 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:37:18.136260157 +0000 UTC m=+25.783945214 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.136357 4771 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.136355 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.136407 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:18.136395711 +0000 UTC m=+25.784080778 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.136486 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.136479 4771 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.136620 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:18.136596607 +0000 UTC m=+25.784281844 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.136504 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.136671 4771 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.136723 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:18.13671279 +0000 UTC m=+25.784398057 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.231401 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.237006 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.237331 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.237384 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.237403 4771 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.237502 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:18.237474947 +0000 UTC m=+25.885160184 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.680479 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.680596 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.680662 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.680498 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.680791 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.680898 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.826428 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.838148 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.841877 4771 scope.go:117] "RemoveContainer" containerID="0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355" Oct 02 09:37:16 crc kubenswrapper[4771]: E1002 09:37:16.842089 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.844680 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.845544 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.858311 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.869489 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.883444 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.896613 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.915234 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.928396 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.947344 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.964184 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:16 crc kubenswrapper[4771]: I1002 09:37:16.985447 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:17 crc kubenswrapper[4771]: I1002 09:37:17.000479 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:17 crc kubenswrapper[4771]: I1002 09:37:17.014895 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:17 crc kubenswrapper[4771]: I1002 09:37:17.032179 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:17 crc kubenswrapper[4771]: I1002 09:37:17.051391 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:17 crc kubenswrapper[4771]: I1002 09:37:17.064504 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:17 crc kubenswrapper[4771]: I1002 09:37:17.077879 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:17 crc kubenswrapper[4771]: I1002 09:37:17.092666 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:17 crc kubenswrapper[4771]: I1002 09:37:17.844802 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2"} Oct 02 09:37:17 crc kubenswrapper[4771]: I1002 09:37:17.863589 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:17 crc kubenswrapper[4771]: I1002 09:37:17.893093 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:17 crc kubenswrapper[4771]: I1002 09:37:17.923285 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:17 crc kubenswrapper[4771]: I1002 09:37:17.964067 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:17 crc kubenswrapper[4771]: I1002 09:37:17.985013 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.000872 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.017170 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.030535 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.042907 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.154299 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.154405 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.154434 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.154465 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.154550 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:37:22.154515563 +0000 UTC m=+29.802200630 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.154601 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.154601 4771 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.154639 4771 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.154622 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.154707 4771 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.154716 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:22.154693028 +0000 UTC m=+29.802378265 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.154738 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:22.154728299 +0000 UTC m=+29.802413556 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.154753 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:22.15474534 +0000 UTC m=+29.802430497 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.255684 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.255847 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.255864 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.255877 4771 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.255948 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:22.255931258 +0000 UTC m=+29.903616325 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.269707 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-5z94m"] Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.270031 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5z94m" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.272272 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.272550 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.274828 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.294167 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.306351 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.316890 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.335280 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.349860 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.356233 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a074140a-f8fe-4e03-90e9-368c0632e369-hosts-file\") pod \"node-resolver-5z94m\" (UID: \"a074140a-f8fe-4e03-90e9-368c0632e369\") " pod="openshift-dns/node-resolver-5z94m" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.356286 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhtxd\" (UniqueName: \"kubernetes.io/projected/a074140a-f8fe-4e03-90e9-368c0632e369-kube-api-access-lhtxd\") pod \"node-resolver-5z94m\" (UID: \"a074140a-f8fe-4e03-90e9-368c0632e369\") " pod="openshift-dns/node-resolver-5z94m" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.365696 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.395249 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.416228 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.445623 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.457663 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a074140a-f8fe-4e03-90e9-368c0632e369-hosts-file\") pod \"node-resolver-5z94m\" (UID: \"a074140a-f8fe-4e03-90e9-368c0632e369\") " pod="openshift-dns/node-resolver-5z94m" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.457742 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhtxd\" (UniqueName: \"kubernetes.io/projected/a074140a-f8fe-4e03-90e9-368c0632e369-kube-api-access-lhtxd\") pod \"node-resolver-5z94m\" (UID: \"a074140a-f8fe-4e03-90e9-368c0632e369\") " pod="openshift-dns/node-resolver-5z94m" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.457837 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a074140a-f8fe-4e03-90e9-368c0632e369-hosts-file\") pod \"node-resolver-5z94m\" (UID: \"a074140a-f8fe-4e03-90e9-368c0632e369\") " pod="openshift-dns/node-resolver-5z94m" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.480608 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.486701 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhtxd\" (UniqueName: \"kubernetes.io/projected/a074140a-f8fe-4e03-90e9-368c0632e369-kube-api-access-lhtxd\") pod \"node-resolver-5z94m\" (UID: \"a074140a-f8fe-4e03-90e9-368c0632e369\") " pod="openshift-dns/node-resolver-5z94m" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.582417 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5z94m" Oct 02 09:37:18 crc kubenswrapper[4771]: W1002 09:37:18.600739 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda074140a_f8fe_4e03_90e9_368c0632e369.slice/crio-dcd2f4722f8f5eb83b818a96a52544a91fd1e507e3a46569ad1b0a2ce867ce36 WatchSource:0}: Error finding container dcd2f4722f8f5eb83b818a96a52544a91fd1e507e3a46569ad1b0a2ce867ce36: Status 404 returned error can't find the container with id dcd2f4722f8f5eb83b818a96a52544a91fd1e507e3a46569ad1b0a2ce867ce36 Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.651497 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-xh7xv"] Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.652028 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-m4mpt"] Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.652345 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.652390 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.656818 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.656818 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.656824 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.657269 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.656903 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.656972 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.657411 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.657067 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.657533 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.657593 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.658260 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-qb5kw"] Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.663282 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: W1002 09:37:18.665074 4771 reflector.go:561] object-"openshift-multus"/"default-cni-sysctl-allowlist": failed to list *v1.ConfigMap: configmaps "default-cni-sysctl-allowlist" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.665108 4771 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"default-cni-sysctl-allowlist\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 09:37:18 crc kubenswrapper[4771]: W1002 09:37:18.665170 4771 reflector.go:561] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": failed to list *v1.Secret: secrets "multus-ancillary-tools-dockercfg-vnmsz" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.665183 4771 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-vnmsz\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"multus-ancillary-tools-dockercfg-vnmsz\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.681156 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.681278 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.681341 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.681384 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.681420 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:18 crc kubenswrapper[4771]: E1002 09:37:18.681458 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.686003 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.708014 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.731850 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.745844 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760020 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b09690eb-cfe8-4891-821e-1194ae265b07-os-release\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760065 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b09690eb-cfe8-4891-821e-1194ae265b07-cni-binary-copy\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760087 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-run-k8s-cni-cncf-io\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760103 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-var-lib-cni-bin\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760141 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b09690eb-cfe8-4891-821e-1194ae265b07-system-cni-dir\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760163 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-cnibin\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760183 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-hostroot\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760203 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbs6q\" (UniqueName: \"kubernetes.io/projected/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-kube-api-access-pbs6q\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760228 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slpkg\" (UniqueName: \"kubernetes.io/projected/52d686b1-d29a-42c4-97ce-e239a0c680b0-kube-api-access-slpkg\") pod \"machine-config-daemon-m4mpt\" (UID: \"52d686b1-d29a-42c4-97ce-e239a0c680b0\") " pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760250 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b09690eb-cfe8-4891-821e-1194ae265b07-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760268 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx86k\" (UniqueName: \"kubernetes.io/projected/b09690eb-cfe8-4891-821e-1194ae265b07-kube-api-access-mx86k\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760287 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-run-netns\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760326 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-var-lib-kubelet\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760360 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-etc-kubernetes\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760383 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/52d686b1-d29a-42c4-97ce-e239a0c680b0-proxy-tls\") pod \"machine-config-daemon-m4mpt\" (UID: \"52d686b1-d29a-42c4-97ce-e239a0c680b0\") " pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760411 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b09690eb-cfe8-4891-821e-1194ae265b07-cnibin\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760433 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-multus-cni-dir\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760455 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-multus-socket-dir-parent\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760482 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b09690eb-cfe8-4891-821e-1194ae265b07-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760503 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-system-cni-dir\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760521 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-os-release\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760539 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-cni-binary-copy\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760582 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-multus-daemon-config\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760642 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-run-multus-certs\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760695 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/52d686b1-d29a-42c4-97ce-e239a0c680b0-rootfs\") pod \"machine-config-daemon-m4mpt\" (UID: \"52d686b1-d29a-42c4-97ce-e239a0c680b0\") " pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760732 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-var-lib-cni-multus\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760764 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-multus-conf-dir\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.760786 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/52d686b1-d29a-42c4-97ce-e239a0c680b0-mcd-auth-proxy-config\") pod \"machine-config-daemon-m4mpt\" (UID: \"52d686b1-d29a-42c4-97ce-e239a0c680b0\") " pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.763397 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.788771 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.810659 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.828690 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.843858 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.848663 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5z94m" event={"ID":"a074140a-f8fe-4e03-90e9-368c0632e369","Type":"ContainerStarted","Data":"45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192"} Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.848711 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5z94m" event={"ID":"a074140a-f8fe-4e03-90e9-368c0632e369","Type":"ContainerStarted","Data":"dcd2f4722f8f5eb83b818a96a52544a91fd1e507e3a46569ad1b0a2ce867ce36"} Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.860038 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861514 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slpkg\" (UniqueName: \"kubernetes.io/projected/52d686b1-d29a-42c4-97ce-e239a0c680b0-kube-api-access-slpkg\") pod \"machine-config-daemon-m4mpt\" (UID: \"52d686b1-d29a-42c4-97ce-e239a0c680b0\") " pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861572 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b09690eb-cfe8-4891-821e-1194ae265b07-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861600 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx86k\" (UniqueName: \"kubernetes.io/projected/b09690eb-cfe8-4891-821e-1194ae265b07-kube-api-access-mx86k\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861628 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-run-netns\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861652 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/52d686b1-d29a-42c4-97ce-e239a0c680b0-proxy-tls\") pod \"machine-config-daemon-m4mpt\" (UID: \"52d686b1-d29a-42c4-97ce-e239a0c680b0\") " pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861671 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-var-lib-kubelet\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861688 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-etc-kubernetes\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861709 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b09690eb-cfe8-4891-821e-1194ae265b07-cnibin\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861728 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-multus-cni-dir\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861730 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-run-netns\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861745 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-multus-socket-dir-parent\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861812 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-etc-kubernetes\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861831 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-os-release\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861848 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b09690eb-cfe8-4891-821e-1194ae265b07-cnibin\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861862 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-cni-binary-copy\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861890 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-multus-daemon-config\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861918 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b09690eb-cfe8-4891-821e-1194ae265b07-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861946 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-system-cni-dir\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861979 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-run-multus-certs\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.862799 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/52d686b1-d29a-42c4-97ce-e239a0c680b0-rootfs\") pod \"machine-config-daemon-m4mpt\" (UID: \"52d686b1-d29a-42c4-97ce-e239a0c680b0\") " pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.862926 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-var-lib-cni-multus\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.863003 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-multus-conf-dir\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.863007 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b09690eb-cfe8-4891-821e-1194ae265b07-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.863056 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/52d686b1-d29a-42c4-97ce-e239a0c680b0-mcd-auth-proxy-config\") pod \"machine-config-daemon-m4mpt\" (UID: \"52d686b1-d29a-42c4-97ce-e239a0c680b0\") " pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.862255 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-run-multus-certs\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.863139 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-multus-daemon-config\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.863157 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b09690eb-cfe8-4891-821e-1194ae265b07-os-release\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.863343 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-cni-binary-copy\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.862168 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-multus-cni-dir\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.861994 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-multus-socket-dir-parent\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.863493 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/52d686b1-d29a-42c4-97ce-e239a0c680b0-rootfs\") pod \"machine-config-daemon-m4mpt\" (UID: \"52d686b1-d29a-42c4-97ce-e239a0c680b0\") " pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.862153 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-var-lib-kubelet\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.862355 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-os-release\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.862208 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-system-cni-dir\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.863825 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-var-lib-cni-multus\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864161 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b09690eb-cfe8-4891-821e-1194ae265b07-cni-binary-copy\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864186 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-multus-conf-dir\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864250 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-run-k8s-cni-cncf-io\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864259 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/52d686b1-d29a-42c4-97ce-e239a0c680b0-mcd-auth-proxy-config\") pod \"machine-config-daemon-m4mpt\" (UID: \"52d686b1-d29a-42c4-97ce-e239a0c680b0\") " pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864300 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-var-lib-cni-bin\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864331 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b09690eb-cfe8-4891-821e-1194ae265b07-os-release\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864421 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-run-k8s-cni-cncf-io\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864482 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-hostroot\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864525 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b09690eb-cfe8-4891-821e-1194ae265b07-system-cni-dir\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864637 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-cnibin\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864672 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbs6q\" (UniqueName: \"kubernetes.io/projected/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-kube-api-access-pbs6q\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864703 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b09690eb-cfe8-4891-821e-1194ae265b07-cni-binary-copy\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864725 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-host-var-lib-cni-bin\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864864 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b09690eb-cfe8-4891-821e-1194ae265b07-system-cni-dir\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864901 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-hostroot\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.864987 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-cnibin\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.868205 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/52d686b1-d29a-42c4-97ce-e239a0c680b0-proxy-tls\") pod \"machine-config-daemon-m4mpt\" (UID: \"52d686b1-d29a-42c4-97ce-e239a0c680b0\") " pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.878089 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.883710 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slpkg\" (UniqueName: \"kubernetes.io/projected/52d686b1-d29a-42c4-97ce-e239a0c680b0-kube-api-access-slpkg\") pod \"machine-config-daemon-m4mpt\" (UID: \"52d686b1-d29a-42c4-97ce-e239a0c680b0\") " pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.885401 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx86k\" (UniqueName: \"kubernetes.io/projected/b09690eb-cfe8-4891-821e-1194ae265b07-kube-api-access-mx86k\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.886841 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbs6q\" (UniqueName: \"kubernetes.io/projected/bb6b65c1-83b7-4b23-abe8-c4255c6e59f8-kube-api-access-pbs6q\") pod \"multus-xh7xv\" (UID: \"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\") " pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.897306 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.927232 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.954395 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.967600 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xh7xv" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.973448 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:18 crc kubenswrapper[4771]: W1002 09:37:18.977741 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb6b65c1_83b7_4b23_abe8_c4255c6e59f8.slice/crio-dc3074310fadb9e3870b920a951bffb1f79daccc5e187d936eb4b01b9125ac61 WatchSource:0}: Error finding container dc3074310fadb9e3870b920a951bffb1f79daccc5e187d936eb4b01b9125ac61: Status 404 returned error can't find the container with id dc3074310fadb9e3870b920a951bffb1f79daccc5e187d936eb4b01b9125ac61 Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.980147 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:37:18 crc kubenswrapper[4771]: I1002 09:37:18.988186 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.005564 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.028624 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.055823 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4lxxg"] Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.056957 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.061076 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.062322 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.062511 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.062673 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.062831 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.063113 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.063420 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.067148 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.087140 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.106841 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.128341 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.144673 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.165273 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.167659 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-log-socket\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.167703 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovnkube-config\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.167723 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovnkube-script-lib\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.167743 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-systemd-units\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.167832 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-etc-openvswitch\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.167856 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-run-netns\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.167931 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-run-ovn-kubernetes\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.167955 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-cni-bin\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.167985 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-kubelet\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.168005 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl2v4\" (UniqueName: \"kubernetes.io/projected/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-kube-api-access-pl2v4\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.168053 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-env-overrides\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.168168 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-slash\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.168201 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-openvswitch\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.168256 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.168289 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-systemd\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.168316 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-node-log\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.168350 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovn-node-metrics-cert\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.168384 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-var-lib-openvswitch\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.168403 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-cni-netd\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.168452 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-ovn\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.180612 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.195954 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.209318 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.221957 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.236703 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.252423 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.269612 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.269886 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.269962 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-systemd\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.269992 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-node-log\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270049 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-var-lib-openvswitch\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270057 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270091 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-cni-netd\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270147 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovn-node-metrics-cert\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270180 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-ovn\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270180 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-systemd\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270151 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-var-lib-openvswitch\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270207 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-log-socket\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270291 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-log-socket\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270311 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-ovn\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270308 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-cni-netd\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270355 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovnkube-config\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270406 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-node-log\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270444 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovnkube-script-lib\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270489 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-systemd-units\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270516 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-etc-openvswitch\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270547 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-run-netns\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270587 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-run-ovn-kubernetes\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270616 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-cni-bin\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270668 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-kubelet\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270698 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl2v4\" (UniqueName: \"kubernetes.io/projected/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-kube-api-access-pl2v4\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270739 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-env-overrides\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270767 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-openvswitch\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270797 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-slash\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270867 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-slash\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270912 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-systemd-units\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270936 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-etc-openvswitch\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270961 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-run-netns\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.270987 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-run-ovn-kubernetes\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.271010 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-cni-bin\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.271033 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-kubelet\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.271158 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovnkube-config\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.271250 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-openvswitch\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.271403 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovnkube-script-lib\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.271592 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-env-overrides\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.274618 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovn-node-metrics-cert\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.289459 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl2v4\" (UniqueName: \"kubernetes.io/projected/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-kube-api-access-pl2v4\") pod \"ovnkube-node-4lxxg\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.293716 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.316963 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.335113 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.347675 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.362203 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.372024 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.379274 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: W1002 09:37:19.385033 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c5d0dc7_d6bf_4c34_aba5_a14dbdbde618.slice/crio-0db0a7b33494b001ac2efb0394aa34a8b70b62735c35c50f0e85d8af47548847 WatchSource:0}: Error finding container 0db0a7b33494b001ac2efb0394aa34a8b70b62735c35c50f0e85d8af47548847: Status 404 returned error can't find the container with id 0db0a7b33494b001ac2efb0394aa34a8b70b62735c35c50f0e85d8af47548847 Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.398642 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.601685 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.603239 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b09690eb-cfe8-4891-821e-1194ae265b07-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qb5kw\" (UID: \"b09690eb-cfe8-4891-821e-1194ae265b07\") " pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.688388 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.852941 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7"} Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.853003 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a"} Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.853018 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"06fff18f9d2181bc8b2edbf04ef4d0e5603599b1aade50f2f34d8aec83910513"} Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.854530 4771 generic.go:334] "Generic (PLEG): container finished" podID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerID="f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68" exitCode=0 Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.854609 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68"} Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.854646 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerStarted","Data":"0db0a7b33494b001ac2efb0394aa34a8b70b62735c35c50f0e85d8af47548847"} Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.856940 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xh7xv" event={"ID":"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8","Type":"ContainerStarted","Data":"a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c"} Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.857001 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xh7xv" event={"ID":"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8","Type":"ContainerStarted","Data":"dc3074310fadb9e3870b920a951bffb1f79daccc5e187d936eb4b01b9125ac61"} Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.866590 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.885511 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.888447 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.901009 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: W1002 09:37:19.906469 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb09690eb_cfe8_4891_821e_1194ae265b07.slice/crio-588987cfc344164227d2eef42860539aff4134c3d78357684bc132ca20c0e449 WatchSource:0}: Error finding container 588987cfc344164227d2eef42860539aff4134c3d78357684bc132ca20c0e449: Status 404 returned error can't find the container with id 588987cfc344164227d2eef42860539aff4134c3d78357684bc132ca20c0e449 Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.925082 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.942563 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.961279 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.980640 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:19 crc kubenswrapper[4771]: I1002 09:37:19.993066 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.007631 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.029975 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.041075 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.051907 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.074036 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.092035 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.110773 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.123218 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.135096 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.154001 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.168692 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.181866 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.197076 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.210914 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.223349 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.239517 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.258974 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.281062 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.296818 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.317712 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.565410 4771 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.568467 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.568505 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.568517 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.568621 4771 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.580427 4771 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.580797 4771 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.582058 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.582106 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.582119 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.582159 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.582172 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:20Z","lastTransitionTime":"2025-10-02T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:20 crc kubenswrapper[4771]: E1002 09:37:20.605223 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.610501 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.610546 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.610603 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.610628 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.610642 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:20Z","lastTransitionTime":"2025-10-02T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:20 crc kubenswrapper[4771]: E1002 09:37:20.622000 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.625423 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.625445 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.625453 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.625468 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.625480 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:20Z","lastTransitionTime":"2025-10-02T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:20 crc kubenswrapper[4771]: E1002 09:37:20.638913 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.642856 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.642898 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.642908 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.642930 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.642942 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:20Z","lastTransitionTime":"2025-10-02T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:20 crc kubenswrapper[4771]: E1002 09:37:20.658500 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.662925 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.662971 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.662982 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.663000 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.663011 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:20Z","lastTransitionTime":"2025-10-02T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:20 crc kubenswrapper[4771]: E1002 09:37:20.677199 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: E1002 09:37:20.677379 4771 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.679696 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.679747 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.679763 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.679785 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.679797 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:20Z","lastTransitionTime":"2025-10-02T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.680288 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.680296 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:20 crc kubenswrapper[4771]: E1002 09:37:20.680455 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.680407 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:20 crc kubenswrapper[4771]: E1002 09:37:20.680546 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:20 crc kubenswrapper[4771]: E1002 09:37:20.680820 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.783372 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.783418 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.783429 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.783473 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.783486 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:20Z","lastTransitionTime":"2025-10-02T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.870879 4771 generic.go:334] "Generic (PLEG): container finished" podID="b09690eb-cfe8-4891-821e-1194ae265b07" containerID="46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590" exitCode=0 Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.870953 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" event={"ID":"b09690eb-cfe8-4891-821e-1194ae265b07","Type":"ContainerDied","Data":"46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590"} Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.871006 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" event={"ID":"b09690eb-cfe8-4891-821e-1194ae265b07","Type":"ContainerStarted","Data":"588987cfc344164227d2eef42860539aff4134c3d78357684bc132ca20c0e449"} Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.875025 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerStarted","Data":"f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392"} Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.875091 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerStarted","Data":"82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086"} Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.875104 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerStarted","Data":"86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3"} Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.875113 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerStarted","Data":"3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2"} Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.886692 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.886750 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.886799 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.886824 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.886840 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:20Z","lastTransitionTime":"2025-10-02T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.897372 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.920318 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.938796 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.956277 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.980532 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.993697 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.993733 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.993742 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.993760 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.993770 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:20Z","lastTransitionTime":"2025-10-02T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:20 crc kubenswrapper[4771]: I1002 09:37:20.999727 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.019423 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.038904 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.052205 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.063162 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.078018 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.092358 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.096077 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.096188 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.096205 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.096229 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.096243 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:21Z","lastTransitionTime":"2025-10-02T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.107579 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.124178 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.199863 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.199926 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.199940 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.199964 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.199979 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:21Z","lastTransitionTime":"2025-10-02T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.303675 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.303726 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.303738 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.303759 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.303770 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:21Z","lastTransitionTime":"2025-10-02T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.406041 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.406093 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.406104 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.406122 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.406147 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:21Z","lastTransitionTime":"2025-10-02T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.513313 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.513828 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.513844 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.513869 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.513886 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:21Z","lastTransitionTime":"2025-10-02T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.617879 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.617943 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.617952 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.617973 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.617985 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:21Z","lastTransitionTime":"2025-10-02T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.685691 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-qfhxl"] Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.691895 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qfhxl" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.694720 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.695395 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.695524 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.695952 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.713769 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.721153 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.721192 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.721203 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.721221 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.721234 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:21Z","lastTransitionTime":"2025-10-02T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.727469 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.739683 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.749923 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.762065 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.775360 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.788237 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.799016 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f29540bb-d982-4a75-9ee6-e24242514b05-serviceca\") pod \"node-ca-qfhxl\" (UID: \"f29540bb-d982-4a75-9ee6-e24242514b05\") " pod="openshift-image-registry/node-ca-qfhxl" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.799100 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f29540bb-d982-4a75-9ee6-e24242514b05-host\") pod \"node-ca-qfhxl\" (UID: \"f29540bb-d982-4a75-9ee6-e24242514b05\") " pod="openshift-image-registry/node-ca-qfhxl" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.799289 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw4hv\" (UniqueName: \"kubernetes.io/projected/f29540bb-d982-4a75-9ee6-e24242514b05-kube-api-access-nw4hv\") pod \"node-ca-qfhxl\" (UID: \"f29540bb-d982-4a75-9ee6-e24242514b05\") " pod="openshift-image-registry/node-ca-qfhxl" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.809967 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.824355 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.824400 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.824416 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.824436 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.824448 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:21Z","lastTransitionTime":"2025-10-02T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.824498 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.838176 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.851436 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.865089 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.880625 4771 generic.go:334] "Generic (PLEG): container finished" podID="b09690eb-cfe8-4891-821e-1194ae265b07" containerID="1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875" exitCode=0 Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.880694 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" event={"ID":"b09690eb-cfe8-4891-821e-1194ae265b07","Type":"ContainerDied","Data":"1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875"} Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.884912 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.886452 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerStarted","Data":"b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802"} Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.886519 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerStarted","Data":"7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85"} Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.901267 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f29540bb-d982-4a75-9ee6-e24242514b05-serviceca\") pod \"node-ca-qfhxl\" (UID: \"f29540bb-d982-4a75-9ee6-e24242514b05\") " pod="openshift-image-registry/node-ca-qfhxl" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.901310 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f29540bb-d982-4a75-9ee6-e24242514b05-host\") pod \"node-ca-qfhxl\" (UID: \"f29540bb-d982-4a75-9ee6-e24242514b05\") " pod="openshift-image-registry/node-ca-qfhxl" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.901346 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw4hv\" (UniqueName: \"kubernetes.io/projected/f29540bb-d982-4a75-9ee6-e24242514b05-kube-api-access-nw4hv\") pod \"node-ca-qfhxl\" (UID: \"f29540bb-d982-4a75-9ee6-e24242514b05\") " pod="openshift-image-registry/node-ca-qfhxl" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.901495 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f29540bb-d982-4a75-9ee6-e24242514b05-host\") pod \"node-ca-qfhxl\" (UID: \"f29540bb-d982-4a75-9ee6-e24242514b05\") " pod="openshift-image-registry/node-ca-qfhxl" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.902707 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f29540bb-d982-4a75-9ee6-e24242514b05-serviceca\") pod \"node-ca-qfhxl\" (UID: \"f29540bb-d982-4a75-9ee6-e24242514b05\") " pod="openshift-image-registry/node-ca-qfhxl" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.907608 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.918183 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.928665 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.928711 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.928725 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.928747 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.928767 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:21Z","lastTransitionTime":"2025-10-02T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.930145 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw4hv\" (UniqueName: \"kubernetes.io/projected/f29540bb-d982-4a75-9ee6-e24242514b05-kube-api-access-nw4hv\") pod \"node-ca-qfhxl\" (UID: \"f29540bb-d982-4a75-9ee6-e24242514b05\") " pod="openshift-image-registry/node-ca-qfhxl" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.932717 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.949963 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.973536 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:21 crc kubenswrapper[4771]: I1002 09:37:21.995019 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.008706 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qfhxl" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.008819 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.021192 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.032570 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.032617 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.032629 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.032657 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.032669 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:22Z","lastTransitionTime":"2025-10-02T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.038926 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.054893 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.089509 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.113053 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.136638 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.150355 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.150392 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.150401 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.150418 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.150430 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:22Z","lastTransitionTime":"2025-10-02T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.159486 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.176629 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.190089 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.203486 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.207808 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.207967 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.208024 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:37:30.207987946 +0000 UTC m=+37.855673013 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.208084 4771 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.208089 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.208178 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.208205 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:30.208174752 +0000 UTC m=+37.855859819 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.208381 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.208404 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.208417 4771 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.208453 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:30.208445509 +0000 UTC m=+37.856130576 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.208494 4771 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.208522 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:30.208516191 +0000 UTC m=+37.856201258 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.253690 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.253753 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.253768 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.253787 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.253800 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:22Z","lastTransitionTime":"2025-10-02T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.309028 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.309237 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.309259 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.309274 4771 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.309341 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:30.309324009 +0000 UTC m=+37.957009086 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.358091 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.358139 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.358148 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.358164 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.358177 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:22Z","lastTransitionTime":"2025-10-02T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.460254 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.460296 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.460307 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.460328 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.460340 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:22Z","lastTransitionTime":"2025-10-02T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.563695 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.563755 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.563773 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.563799 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.563819 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:22Z","lastTransitionTime":"2025-10-02T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.666631 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.666680 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.666697 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.666721 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.666741 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:22Z","lastTransitionTime":"2025-10-02T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.680845 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.680845 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.681026 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.681217 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.681351 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:22 crc kubenswrapper[4771]: E1002 09:37:22.681482 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.769736 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.769774 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.769783 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.769800 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.769814 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:22Z","lastTransitionTime":"2025-10-02T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.872531 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.872590 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.872610 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.872632 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.872646 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:22Z","lastTransitionTime":"2025-10-02T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.894385 4771 generic.go:334] "Generic (PLEG): container finished" podID="b09690eb-cfe8-4891-821e-1194ae265b07" containerID="5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4" exitCode=0 Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.894503 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" event={"ID":"b09690eb-cfe8-4891-821e-1194ae265b07","Type":"ContainerDied","Data":"5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4"} Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.896425 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qfhxl" event={"ID":"f29540bb-d982-4a75-9ee6-e24242514b05","Type":"ContainerStarted","Data":"2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f"} Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.896495 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qfhxl" event={"ID":"f29540bb-d982-4a75-9ee6-e24242514b05","Type":"ContainerStarted","Data":"b39c4ad561e5ebba4557899289036cf996d8d6703deeced1d7b0771255020ac0"} Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.917983 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.933684 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.948897 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.971163 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.975792 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.975827 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.975836 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.975849 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.975859 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:22Z","lastTransitionTime":"2025-10-02T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:22 crc kubenswrapper[4771]: I1002 09:37:22.989249 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.010280 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.024038 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.039887 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.069841 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.078449 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.078496 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.078509 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.078532 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.078547 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:23Z","lastTransitionTime":"2025-10-02T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.087695 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.103941 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.116712 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.130594 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.143603 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.154690 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.173437 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.183247 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.183281 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.183295 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.183313 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.183326 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:23Z","lastTransitionTime":"2025-10-02T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.187620 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.199573 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.212613 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.225030 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.239257 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.275419 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.286121 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.286199 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.286209 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.286228 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.286239 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:23Z","lastTransitionTime":"2025-10-02T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.291460 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.310924 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.320822 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.332330 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.348121 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.359659 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.371338 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.380384 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.389087 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.389113 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.389137 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.389152 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.389161 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:23Z","lastTransitionTime":"2025-10-02T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.492380 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.492466 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.492490 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.492528 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.492547 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:23Z","lastTransitionTime":"2025-10-02T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.595556 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.595625 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.595650 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.595682 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.595708 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:23Z","lastTransitionTime":"2025-10-02T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.698555 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.698616 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.698631 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.698649 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.698662 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:23Z","lastTransitionTime":"2025-10-02T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.703853 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.719153 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.736313 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.754630 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.770154 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.784677 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.801874 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.801928 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.801938 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.801957 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.801968 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:23Z","lastTransitionTime":"2025-10-02T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.804920 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.816814 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.829572 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.843787 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.861339 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.877584 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.895992 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.903827 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.903868 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.903885 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.903904 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.903917 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:23Z","lastTransitionTime":"2025-10-02T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.905110 4771 generic.go:334] "Generic (PLEG): container finished" podID="b09690eb-cfe8-4891-821e-1194ae265b07" containerID="aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d" exitCode=0 Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.905237 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" event={"ID":"b09690eb-cfe8-4891-821e-1194ae265b07","Type":"ContainerDied","Data":"aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d"} Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.915925 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.923121 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerStarted","Data":"6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa"} Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.931635 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.947508 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.962534 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:23 crc kubenswrapper[4771]: I1002 09:37:23.980999 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.001172 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.009500 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.009528 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.009537 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.009551 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.009560 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:24Z","lastTransitionTime":"2025-10-02T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.022100 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.037625 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.058625 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.069859 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.080421 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.094591 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.109891 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.112000 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.112031 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.112039 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.112054 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.112064 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:24Z","lastTransitionTime":"2025-10-02T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.121740 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.134963 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.161444 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.200689 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.214826 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.214875 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.214888 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.214907 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.214918 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:24Z","lastTransitionTime":"2025-10-02T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.318273 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.318325 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.318341 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.318364 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.318381 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:24Z","lastTransitionTime":"2025-10-02T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.420756 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.420806 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.420818 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.420835 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.420847 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:24Z","lastTransitionTime":"2025-10-02T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.524035 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.524075 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.524087 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.524105 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.524117 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:24Z","lastTransitionTime":"2025-10-02T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.627450 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.627521 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.627537 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.627566 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.627589 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:24Z","lastTransitionTime":"2025-10-02T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.680487 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.680586 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.680495 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:24 crc kubenswrapper[4771]: E1002 09:37:24.680691 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:24 crc kubenswrapper[4771]: E1002 09:37:24.680832 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:24 crc kubenswrapper[4771]: E1002 09:37:24.680969 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.730075 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.730182 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.730211 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.730246 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.730272 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:24Z","lastTransitionTime":"2025-10-02T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.835154 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.835219 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.835228 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.835252 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.835263 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:24Z","lastTransitionTime":"2025-10-02T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.930089 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" event={"ID":"b09690eb-cfe8-4891-821e-1194ae265b07","Type":"ContainerStarted","Data":"b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97"} Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.938795 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.938856 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.938881 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.938905 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.938927 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:24Z","lastTransitionTime":"2025-10-02T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.947809 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.960289 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.977669 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:24 crc kubenswrapper[4771]: I1002 09:37:24.991433 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.004325 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.015980 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.029673 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.042403 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.042450 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.042460 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.042477 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.042490 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:25Z","lastTransitionTime":"2025-10-02T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.047779 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.064818 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.092762 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.108010 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.125631 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.139010 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.144571 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.144603 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.144611 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.144627 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.144637 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:25Z","lastTransitionTime":"2025-10-02T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.161276 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.173723 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.246979 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.247041 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.247058 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.247081 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.247097 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:25Z","lastTransitionTime":"2025-10-02T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.349887 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.349986 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.350003 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.350048 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.350065 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:25Z","lastTransitionTime":"2025-10-02T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.453069 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.453176 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.453203 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.453236 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.453260 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:25Z","lastTransitionTime":"2025-10-02T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.556226 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.556274 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.556288 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.556308 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.556322 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:25Z","lastTransitionTime":"2025-10-02T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.659447 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.659515 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.659536 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.659657 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.659679 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:25Z","lastTransitionTime":"2025-10-02T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.762590 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.762646 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.762660 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.762679 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.762707 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:25Z","lastTransitionTime":"2025-10-02T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.866260 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.866317 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.866335 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.866361 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.866378 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:25Z","lastTransitionTime":"2025-10-02T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.940118 4771 generic.go:334] "Generic (PLEG): container finished" podID="b09690eb-cfe8-4891-821e-1194ae265b07" containerID="b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97" exitCode=0 Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.940225 4771 generic.go:334] "Generic (PLEG): container finished" podID="b09690eb-cfe8-4891-821e-1194ae265b07" containerID="bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85" exitCode=0 Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.940302 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" event={"ID":"b09690eb-cfe8-4891-821e-1194ae265b07","Type":"ContainerDied","Data":"b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97"} Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.940359 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" event={"ID":"b09690eb-cfe8-4891-821e-1194ae265b07","Type":"ContainerDied","Data":"bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85"} Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.950097 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerStarted","Data":"716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0"} Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.950456 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.961282 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.969428 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.969480 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.969491 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.969512 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.969526 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:25Z","lastTransitionTime":"2025-10-02T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.976267 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:25 crc kubenswrapper[4771]: I1002 09:37:25.988686 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.001355 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.016522 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.037271 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.038380 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.058435 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.072784 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.072814 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.072823 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.072838 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.072849 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:26Z","lastTransitionTime":"2025-10-02T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.080521 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.102568 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.119971 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.134769 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.151183 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.170550 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.177688 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.177730 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.177742 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.177759 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.177771 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:26Z","lastTransitionTime":"2025-10-02T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.191387 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.205871 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.221180 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.235042 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.247176 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.260430 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.273046 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.280098 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.280147 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.280157 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.280175 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.280195 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:26Z","lastTransitionTime":"2025-10-02T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.284464 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.297973 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.317708 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.331347 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.348546 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.375834 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.382881 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.382951 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.383010 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.383045 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.383059 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:26Z","lastTransitionTime":"2025-10-02T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.437562 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.454462 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.470820 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.485507 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.485555 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.485565 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.485584 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.485596 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:26Z","lastTransitionTime":"2025-10-02T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.492308 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.589061 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.589176 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.589198 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.589226 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.589245 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:26Z","lastTransitionTime":"2025-10-02T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.681036 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.681036 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:26 crc kubenswrapper[4771]: E1002 09:37:26.681232 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.681411 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:26 crc kubenswrapper[4771]: E1002 09:37:26.681568 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:26 crc kubenswrapper[4771]: E1002 09:37:26.681750 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.692402 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.692462 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.692476 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.692495 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.692507 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:26Z","lastTransitionTime":"2025-10-02T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.795159 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.795205 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.795219 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.795239 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.795254 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:26Z","lastTransitionTime":"2025-10-02T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.906145 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.906195 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.906206 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.906223 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.906234 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:26Z","lastTransitionTime":"2025-10-02T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.960070 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" event={"ID":"b09690eb-cfe8-4891-821e-1194ae265b07","Type":"ContainerStarted","Data":"384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9"} Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.960158 4771 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.960663 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.984567 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:26 crc kubenswrapper[4771]: I1002 09:37:26.986202 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.001681 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.010193 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.010248 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.010260 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.010279 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.010292 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:27Z","lastTransitionTime":"2025-10-02T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.023944 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.048755 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.074616 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.095036 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.113599 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.113659 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.113671 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.113694 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.113710 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:27Z","lastTransitionTime":"2025-10-02T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.116079 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.133588 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.158615 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.169973 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.185929 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.200343 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.213897 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.215751 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.215798 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.215809 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.215827 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.215840 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:27Z","lastTransitionTime":"2025-10-02T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.229888 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.243551 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.255119 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.270465 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.286518 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.312119 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.318423 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.318468 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.318479 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.318520 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.318535 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:27Z","lastTransitionTime":"2025-10-02T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.331180 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.346041 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.359784 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.375252 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.389880 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.405573 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.421239 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.421288 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.421297 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.421317 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.421329 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:27Z","lastTransitionTime":"2025-10-02T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.431211 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.445263 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.466310 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.477421 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.496374 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.524177 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.524214 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.524223 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.524238 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.524248 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:27Z","lastTransitionTime":"2025-10-02T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.626942 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.626984 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.626992 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.627010 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.627020 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:27Z","lastTransitionTime":"2025-10-02T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.729725 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.729866 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.729882 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.729898 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.729919 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:27Z","lastTransitionTime":"2025-10-02T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.831789 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.831852 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.831865 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.831882 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.831895 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:27Z","lastTransitionTime":"2025-10-02T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.934853 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.935314 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.935338 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.935362 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.935384 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:27Z","lastTransitionTime":"2025-10-02T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:27 crc kubenswrapper[4771]: I1002 09:37:27.965912 4771 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.038511 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.038552 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.038562 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.038579 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.038591 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:28Z","lastTransitionTime":"2025-10-02T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.141541 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.141577 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.141586 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.141601 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.141612 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:28Z","lastTransitionTime":"2025-10-02T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.244674 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.244736 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.244759 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.244789 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.244816 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:28Z","lastTransitionTime":"2025-10-02T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.347670 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.347745 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.347769 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.347797 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.347818 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:28Z","lastTransitionTime":"2025-10-02T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.450508 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.450563 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.450573 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.450593 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.450607 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:28Z","lastTransitionTime":"2025-10-02T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.553287 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.553338 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.553351 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.553370 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.553386 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:28Z","lastTransitionTime":"2025-10-02T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.661492 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.661555 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.661579 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.661604 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.661623 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:28Z","lastTransitionTime":"2025-10-02T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.680490 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:28 crc kubenswrapper[4771]: E1002 09:37:28.680619 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.680955 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.681249 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:28 crc kubenswrapper[4771]: E1002 09:37:28.681438 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.681487 4771 scope.go:117] "RemoveContainer" containerID="0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355" Oct 02 09:37:28 crc kubenswrapper[4771]: E1002 09:37:28.681663 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.764828 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.764902 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.764917 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.764937 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.764949 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:28Z","lastTransitionTime":"2025-10-02T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.867509 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.867567 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.867579 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.867594 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.867609 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:28Z","lastTransitionTime":"2025-10-02T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.970499 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.970566 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.970579 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.970598 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.970611 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:28Z","lastTransitionTime":"2025-10-02T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.973313 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.975902 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407"} Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.976014 4771 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 09:37:28 crc kubenswrapper[4771]: I1002 09:37:28.994595 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.012600 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.027834 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.054683 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.068465 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.073090 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.073175 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.073190 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.073209 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.073222 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:29Z","lastTransitionTime":"2025-10-02T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.079535 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.098529 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.110146 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.121008 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.130859 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.142283 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.151266 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.161706 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.175834 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.175885 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.175900 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.175918 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.175930 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:29Z","lastTransitionTime":"2025-10-02T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.177682 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.202657 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.279599 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.279685 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.279703 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.279732 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.279753 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:29Z","lastTransitionTime":"2025-10-02T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.383014 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.383049 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.383059 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.383085 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.383101 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:29Z","lastTransitionTime":"2025-10-02T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.486702 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.486747 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.486761 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.486779 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.486793 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:29Z","lastTransitionTime":"2025-10-02T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.590034 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.590083 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.590097 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.590114 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.590439 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:29Z","lastTransitionTime":"2025-10-02T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.691941 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.691982 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.691993 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.692008 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.692019 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:29Z","lastTransitionTime":"2025-10-02T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.793864 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.793990 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.794002 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.794017 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.794025 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:29Z","lastTransitionTime":"2025-10-02T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.896472 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.896829 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.896899 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.897034 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.897286 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:29Z","lastTransitionTime":"2025-10-02T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.981451 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/0.log" Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.984971 4771 generic.go:334] "Generic (PLEG): container finished" podID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerID="716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0" exitCode=1 Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.985026 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0"} Oct 02 09:37:29 crc kubenswrapper[4771]: I1002 09:37:29.985848 4771 scope.go:117] "RemoveContainer" containerID="716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.000614 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.000880 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.000893 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.000910 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.000920 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:30Z","lastTransitionTime":"2025-10-02T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.006715 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.029614 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.046777 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.066019 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.080003 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.092414 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.103695 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.103977 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.104065 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.104167 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.104265 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:30Z","lastTransitionTime":"2025-10-02T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.106318 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.123599 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.140477 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.161861 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.185923 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.202754 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.206332 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.206437 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.206498 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.206575 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.206666 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:30Z","lastTransitionTime":"2025-10-02T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.217057 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.228925 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.258106 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:28Z\\\",\\\"message\\\":\\\"bernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:37:28.973306 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:37:28.973599 6111 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:37:28.974698 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 09:37:28.974814 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 09:37:28.974871 6111 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 09:37:28.974901 6111 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 09:37:28.974955 6111 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 09:37:28.974898 6111 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 09:37:28.974881 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 09:37:28.975049 6111 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 09:37:28.975084 6111 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 09:37:28.975417 6111 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 09:37:28.975587 6111 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:30Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.297884 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.298390 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:37:46.29835533 +0000 UTC m=+53.946040407 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.298453 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.298489 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.298523 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.298698 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.298718 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.298719 4771 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.298782 4771 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.298855 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:46.298833774 +0000 UTC m=+53.946518851 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.298731 4771 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.298925 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:46.298886665 +0000 UTC m=+53.946571762 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.299031 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:46.299003288 +0000 UTC m=+53.946688535 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.310032 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.310073 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.310082 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.310098 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.310109 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:30Z","lastTransitionTime":"2025-10-02T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.400003 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.400213 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.400230 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.400242 4771 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.400292 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:46.400278446 +0000 UTC m=+54.047963513 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.413561 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.413605 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.413620 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.413641 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.413655 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:30Z","lastTransitionTime":"2025-10-02T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.516833 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.516897 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.516917 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.516949 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.517016 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:30Z","lastTransitionTime":"2025-10-02T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.621006 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.621049 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.621060 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.621074 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.621085 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:30Z","lastTransitionTime":"2025-10-02T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.680371 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.680460 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.680384 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.680626 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.680748 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:30 crc kubenswrapper[4771]: E1002 09:37:30.680881 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.723904 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.723949 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.723981 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.723998 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.724043 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:30Z","lastTransitionTime":"2025-10-02T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.827298 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.827348 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.827362 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.827380 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.827395 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:30Z","lastTransitionTime":"2025-10-02T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.931370 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.931429 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.931451 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.931484 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.931508 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:30Z","lastTransitionTime":"2025-10-02T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.991371 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/0.log" Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.993904 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerStarted","Data":"208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee"} Oct 02 09:37:30 crc kubenswrapper[4771]: I1002 09:37:30.994085 4771 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.008298 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.016062 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.016105 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.016115 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.016145 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.016156 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.019277 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: E1002 09:37:31.033167 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.034233 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.038907 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.038965 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.038978 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.039001 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.039016 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.050224 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: E1002 09:37:31.053557 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.058232 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.058285 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.058298 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.058317 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.058330 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.063784 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: E1002 09:37:31.070941 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.075462 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.075499 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.075512 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.075529 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.075540 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.079561 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: E1002 09:37:31.087636 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.091400 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.091465 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.091478 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.091500 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.091514 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.101307 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: E1002 09:37:31.104252 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: E1002 09:37:31.104419 4771 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.106140 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.106168 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.106176 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.106193 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.106207 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.115283 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.130269 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.148978 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.164219 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.186520 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.197099 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr"] Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.197554 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.198955 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.199110 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.209285 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.209327 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.209339 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.209354 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.209364 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.211491 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.233388 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.253976 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:28Z\\\",\\\"message\\\":\\\"bernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:37:28.973306 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:37:28.973599 6111 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:37:28.974698 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 09:37:28.974814 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 09:37:28.974871 6111 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 09:37:28.974901 6111 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 09:37:28.974955 6111 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 09:37:28.974898 6111 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 09:37:28.974881 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 09:37:28.975049 6111 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 09:37:28.975084 6111 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 09:37:28.975417 6111 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 09:37:28.975587 6111 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.270063 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.284712 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.297105 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.309379 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3d6fc451-a597-4cf5-99b9-8161888e8fcd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wf5gr\" (UID: \"3d6fc451-a597-4cf5-99b9-8161888e8fcd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.309774 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3d6fc451-a597-4cf5-99b9-8161888e8fcd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wf5gr\" (UID: \"3d6fc451-a597-4cf5-99b9-8161888e8fcd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.309906 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3d6fc451-a597-4cf5-99b9-8161888e8fcd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wf5gr\" (UID: \"3d6fc451-a597-4cf5-99b9-8161888e8fcd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.310008 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsqb7\" (UniqueName: \"kubernetes.io/projected/3d6fc451-a597-4cf5-99b9-8161888e8fcd-kube-api-access-xsqb7\") pod \"ovnkube-control-plane-749d76644c-wf5gr\" (UID: \"3d6fc451-a597-4cf5-99b9-8161888e8fcd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.312295 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.312425 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.312517 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.312601 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.312621 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.312690 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.330182 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.352776 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.381265 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.397463 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.410813 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3d6fc451-a597-4cf5-99b9-8161888e8fcd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wf5gr\" (UID: \"3d6fc451-a597-4cf5-99b9-8161888e8fcd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.410889 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsqb7\" (UniqueName: \"kubernetes.io/projected/3d6fc451-a597-4cf5-99b9-8161888e8fcd-kube-api-access-xsqb7\") pod \"ovnkube-control-plane-749d76644c-wf5gr\" (UID: \"3d6fc451-a597-4cf5-99b9-8161888e8fcd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.410918 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3d6fc451-a597-4cf5-99b9-8161888e8fcd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wf5gr\" (UID: \"3d6fc451-a597-4cf5-99b9-8161888e8fcd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.410967 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3d6fc451-a597-4cf5-99b9-8161888e8fcd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wf5gr\" (UID: \"3d6fc451-a597-4cf5-99b9-8161888e8fcd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.411699 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3d6fc451-a597-4cf5-99b9-8161888e8fcd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wf5gr\" (UID: \"3d6fc451-a597-4cf5-99b9-8161888e8fcd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.411734 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3d6fc451-a597-4cf5-99b9-8161888e8fcd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wf5gr\" (UID: \"3d6fc451-a597-4cf5-99b9-8161888e8fcd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.414711 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.414738 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.414748 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.414767 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.414778 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.417155 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3d6fc451-a597-4cf5-99b9-8161888e8fcd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wf5gr\" (UID: \"3d6fc451-a597-4cf5-99b9-8161888e8fcd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.424964 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:28Z\\\",\\\"message\\\":\\\"bernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:37:28.973306 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:37:28.973599 6111 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:37:28.974698 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 09:37:28.974814 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 09:37:28.974871 6111 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 09:37:28.974901 6111 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 09:37:28.974955 6111 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 09:37:28.974898 6111 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 09:37:28.974881 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 09:37:28.975049 6111 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 09:37:28.975084 6111 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 09:37:28.975417 6111 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 09:37:28.975587 6111 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.428038 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsqb7\" (UniqueName: \"kubernetes.io/projected/3d6fc451-a597-4cf5-99b9-8161888e8fcd-kube-api-access-xsqb7\") pod \"ovnkube-control-plane-749d76644c-wf5gr\" (UID: \"3d6fc451-a597-4cf5-99b9-8161888e8fcd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.440053 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.455695 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.470344 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.486080 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.507594 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.511812 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.517356 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.517385 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.517393 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.517408 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.517418 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: W1002 09:37:31.530768 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d6fc451_a597_4cf5_99b9_8161888e8fcd.slice/crio-ed536f06bb2032d7e4b3838d5aae4d29d7a1432db25342d7e691e4eafb043d9d WatchSource:0}: Error finding container ed536f06bb2032d7e4b3838d5aae4d29d7a1432db25342d7e691e4eafb043d9d: Status 404 returned error can't find the container with id ed536f06bb2032d7e4b3838d5aae4d29d7a1432db25342d7e691e4eafb043d9d Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.534528 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.557641 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.619993 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.620039 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.620060 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.620079 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.620089 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.723688 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.723749 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.723762 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.723785 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.723799 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.827066 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.827098 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.827108 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.827144 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.827165 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.937493 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.937556 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.937567 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.937589 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.937610 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:31Z","lastTransitionTime":"2025-10-02T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:31 crc kubenswrapper[4771]: I1002 09:37:31.999116 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/1.log" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.000037 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/0.log" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.003175 4771 generic.go:334] "Generic (PLEG): container finished" podID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerID="208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee" exitCode=1 Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.003250 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.003297 4771 scope.go:117] "RemoveContainer" containerID="716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.004516 4771 scope.go:117] "RemoveContainer" containerID="208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee" Oct 02 09:37:32 crc kubenswrapper[4771]: E1002 09:37:32.004796 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.005308 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" event={"ID":"3d6fc451-a597-4cf5-99b9-8161888e8fcd","Type":"ContainerStarted","Data":"65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.005337 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" event={"ID":"3d6fc451-a597-4cf5-99b9-8161888e8fcd","Type":"ContainerStarted","Data":"a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.005346 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" event={"ID":"3d6fc451-a597-4cf5-99b9-8161888e8fcd","Type":"ContainerStarted","Data":"ed536f06bb2032d7e4b3838d5aae4d29d7a1432db25342d7e691e4eafb043d9d"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.020380 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.034511 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.042632 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.042678 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.042692 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.042713 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.042730 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:32Z","lastTransitionTime":"2025-10-02T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.048917 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.064405 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.078964 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.099624 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.115057 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.136857 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.145236 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.145287 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.145299 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.145320 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.145342 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:32Z","lastTransitionTime":"2025-10-02T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.159341 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.185891 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.211930 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.231301 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.247071 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.248718 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.248832 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.248897 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.248971 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.249040 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:32Z","lastTransitionTime":"2025-10-02T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.269743 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:28Z\\\",\\\"message\\\":\\\"bernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:37:28.973306 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:37:28.973599 6111 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:37:28.974698 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 09:37:28.974814 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 09:37:28.974871 6111 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 09:37:28.974901 6111 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 09:37:28.974955 6111 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 09:37:28.974898 6111 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 09:37:28.974881 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 09:37:28.975049 6111 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 09:37:28.975084 6111 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 09:37:28.975417 6111 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 09:37:28.975587 6111 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"points{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:37:31.669786 6287 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669793 6287 services_controller.go:445] Built service openshift-marketplace/redhat-marketplace LB template configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669805 6287 services_controller.go:451] Built service openshift-marketplace/redhat-marketplace cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-marketplace_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 09:37:31.669822 6287 services_controller.go:452] Built service openshift-marketplace/redhat-marketplace per-node LB for\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.285553 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.298710 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.324033 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.339474 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.351456 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.351774 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.351852 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.351937 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.352003 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:32Z","lastTransitionTime":"2025-10-02T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.354956 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.370714 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.386640 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.403844 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.428497 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.447751 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.455206 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.455274 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.455292 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.455314 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.455326 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:32Z","lastTransitionTime":"2025-10-02T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.477252 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:28Z\\\",\\\"message\\\":\\\"bernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:37:28.973306 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:37:28.973599 6111 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:37:28.974698 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 09:37:28.974814 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 09:37:28.974871 6111 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 09:37:28.974901 6111 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 09:37:28.974955 6111 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 09:37:28.974898 6111 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 09:37:28.974881 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 09:37:28.975049 6111 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 09:37:28.975084 6111 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 09:37:28.975417 6111 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 09:37:28.975587 6111 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"points{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:37:31.669786 6287 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669793 6287 services_controller.go:445] Built service openshift-marketplace/redhat-marketplace LB template configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669805 6287 services_controller.go:451] Built service openshift-marketplace/redhat-marketplace cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-marketplace_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 09:37:31.669822 6287 services_controller.go:452] Built service openshift-marketplace/redhat-marketplace per-node LB for\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.494002 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.507758 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.527360 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.540477 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.557242 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.558316 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.558371 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.558384 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.558432 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.558444 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:32Z","lastTransitionTime":"2025-10-02T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.574244 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.587542 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.661217 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.661268 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.661279 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.661299 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.661313 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:32Z","lastTransitionTime":"2025-10-02T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.680988 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.681055 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:32 crc kubenswrapper[4771]: E1002 09:37:32.681160 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.681211 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:32 crc kubenswrapper[4771]: E1002 09:37:32.681469 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:32 crc kubenswrapper[4771]: E1002 09:37:32.681569 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.708496 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-zp7kh"] Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.708995 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:32 crc kubenswrapper[4771]: E1002 09:37:32.709061 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.723769 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.742332 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.761237 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.764003 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.764061 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.764072 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.764091 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.764106 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:32Z","lastTransitionTime":"2025-10-02T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.783287 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.799828 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.815542 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.826942 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.827017 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bfj5\" (UniqueName: \"kubernetes.io/projected/20d5431d-4595-4b96-a8b8-8953e3dffb53-kube-api-access-8bfj5\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.835111 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.850436 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.867550 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.867626 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.867644 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.867671 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.867692 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:32Z","lastTransitionTime":"2025-10-02T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.877632 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:28Z\\\",\\\"message\\\":\\\"bernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:37:28.973306 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:37:28.973599 6111 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:37:28.974698 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 09:37:28.974814 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 09:37:28.974871 6111 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 09:37:28.974901 6111 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 09:37:28.974955 6111 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 09:37:28.974898 6111 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 09:37:28.974881 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 09:37:28.975049 6111 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 09:37:28.975084 6111 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 09:37:28.975417 6111 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 09:37:28.975587 6111 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"points{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:37:31.669786 6287 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669793 6287 services_controller.go:445] Built service openshift-marketplace/redhat-marketplace LB template configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669805 6287 services_controller.go:451] Built service openshift-marketplace/redhat-marketplace cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-marketplace_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 09:37:31.669822 6287 services_controller.go:452] Built service openshift-marketplace/redhat-marketplace per-node LB for\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.899306 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.918565 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.928564 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bfj5\" (UniqueName: \"kubernetes.io/projected/20d5431d-4595-4b96-a8b8-8953e3dffb53-kube-api-access-8bfj5\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.928648 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:32 crc kubenswrapper[4771]: E1002 09:37:32.928822 4771 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:37:32 crc kubenswrapper[4771]: E1002 09:37:32.928927 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs podName:20d5431d-4595-4b96-a8b8-8953e3dffb53 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:33.428902195 +0000 UTC m=+41.076587262 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs") pod "network-metrics-daemon-zp7kh" (UID: "20d5431d-4595-4b96-a8b8-8953e3dffb53") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.931270 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.947856 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bfj5\" (UniqueName: \"kubernetes.io/projected/20d5431d-4595-4b96-a8b8-8953e3dffb53-kube-api-access-8bfj5\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.949949 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.963006 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.971086 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.971183 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.971200 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.971223 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.971239 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:32Z","lastTransitionTime":"2025-10-02T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.977462 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:32 crc kubenswrapper[4771]: I1002 09:37:32.987228 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.000922 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.010901 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/1.log" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.074657 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.074719 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.074737 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.074768 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.074784 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:33Z","lastTransitionTime":"2025-10-02T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.178266 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.178309 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.178321 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.178338 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.178351 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:33Z","lastTransitionTime":"2025-10-02T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.281755 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.281800 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.281816 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.281833 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.281845 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:33Z","lastTransitionTime":"2025-10-02T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.385158 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.385212 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.385223 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.385241 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.385252 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:33Z","lastTransitionTime":"2025-10-02T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.435082 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:33 crc kubenswrapper[4771]: E1002 09:37:33.435327 4771 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:37:33 crc kubenswrapper[4771]: E1002 09:37:33.435456 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs podName:20d5431d-4595-4b96-a8b8-8953e3dffb53 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:34.435430449 +0000 UTC m=+42.083115516 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs") pod "network-metrics-daemon-zp7kh" (UID: "20d5431d-4595-4b96-a8b8-8953e3dffb53") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.487768 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.487817 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.487826 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.487851 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.487862 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:33Z","lastTransitionTime":"2025-10-02T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.590255 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.590667 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.590754 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.590829 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.590903 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:33Z","lastTransitionTime":"2025-10-02T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.693469 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.693534 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.693552 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.693575 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.693587 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:33Z","lastTransitionTime":"2025-10-02T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.696685 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.710391 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.725339 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.736212 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.750403 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.764338 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.782366 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.797009 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.797046 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.797056 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.797073 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.797084 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:33Z","lastTransitionTime":"2025-10-02T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.801483 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.819338 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.842688 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.862353 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.876003 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.901097 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.901927 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.902027 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.902085 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.902179 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.902250 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:33Z","lastTransitionTime":"2025-10-02T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.928546 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:28Z\\\",\\\"message\\\":\\\"bernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:37:28.973306 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:37:28.973599 6111 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:37:28.974698 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 09:37:28.974814 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 09:37:28.974871 6111 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 09:37:28.974901 6111 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 09:37:28.974955 6111 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 09:37:28.974898 6111 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 09:37:28.974881 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 09:37:28.975049 6111 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 09:37:28.975084 6111 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 09:37:28.975417 6111 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 09:37:28.975587 6111 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"points{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:37:31.669786 6287 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669793 6287 services_controller.go:445] Built service openshift-marketplace/redhat-marketplace LB template configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669805 6287 services_controller.go:451] Built service openshift-marketplace/redhat-marketplace cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-marketplace_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 09:37:31.669822 6287 services_controller.go:452] Built service openshift-marketplace/redhat-marketplace per-node LB for\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.944504 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.957870 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:33 crc kubenswrapper[4771]: I1002 09:37:33.972408 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.005048 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.005539 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.005623 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.005761 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.005824 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:34Z","lastTransitionTime":"2025-10-02T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.109530 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.109576 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.109586 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.109602 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.109612 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:34Z","lastTransitionTime":"2025-10-02T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.212310 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.212360 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.212376 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.212400 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.212421 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:34Z","lastTransitionTime":"2025-10-02T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.315519 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.315584 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.315595 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.315616 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.315630 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:34Z","lastTransitionTime":"2025-10-02T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.418255 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.418286 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.418294 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.418308 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.418319 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:34Z","lastTransitionTime":"2025-10-02T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.448358 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:34 crc kubenswrapper[4771]: E1002 09:37:34.448570 4771 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:37:34 crc kubenswrapper[4771]: E1002 09:37:34.448671 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs podName:20d5431d-4595-4b96-a8b8-8953e3dffb53 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:36.448645211 +0000 UTC m=+44.096330488 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs") pod "network-metrics-daemon-zp7kh" (UID: "20d5431d-4595-4b96-a8b8-8953e3dffb53") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.521155 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.521201 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.521213 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.521234 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.521245 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:34Z","lastTransitionTime":"2025-10-02T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.624989 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.625065 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.625078 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.625096 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.625106 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:34Z","lastTransitionTime":"2025-10-02T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.680747 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.680841 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:34 crc kubenswrapper[4771]: E1002 09:37:34.681275 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.680910 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.680867 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:34 crc kubenswrapper[4771]: E1002 09:37:34.681421 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:34 crc kubenswrapper[4771]: E1002 09:37:34.681876 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:34 crc kubenswrapper[4771]: E1002 09:37:34.682059 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.728400 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.728687 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.728809 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.728978 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.729214 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:34Z","lastTransitionTime":"2025-10-02T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.833402 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.833495 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.833520 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.833550 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.833574 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:34Z","lastTransitionTime":"2025-10-02T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.936781 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.936826 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.936835 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.936851 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:34 crc kubenswrapper[4771]: I1002 09:37:34.936861 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:34Z","lastTransitionTime":"2025-10-02T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.039526 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.039576 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.039587 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.039605 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.039616 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:35Z","lastTransitionTime":"2025-10-02T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.143097 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.143162 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.143173 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.143192 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.143206 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:35Z","lastTransitionTime":"2025-10-02T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.245647 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.246345 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.246386 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.246415 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.246434 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:35Z","lastTransitionTime":"2025-10-02T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.349260 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.349353 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.349363 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.349379 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.349392 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:35Z","lastTransitionTime":"2025-10-02T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.452458 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.452499 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.452512 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.452529 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.452543 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:35Z","lastTransitionTime":"2025-10-02T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.555233 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.555286 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.555299 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.555322 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.555337 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:35Z","lastTransitionTime":"2025-10-02T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.658360 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.658514 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.658530 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.658549 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.658560 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:35Z","lastTransitionTime":"2025-10-02T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.761387 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.761434 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.761462 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.761484 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.761496 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:35Z","lastTransitionTime":"2025-10-02T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.864768 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.864817 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.864830 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.864852 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.864864 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:35Z","lastTransitionTime":"2025-10-02T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.967948 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.967992 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.968003 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.968019 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:35 crc kubenswrapper[4771]: I1002 09:37:35.968033 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:35Z","lastTransitionTime":"2025-10-02T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.071578 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.071649 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.071659 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.071681 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.071693 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:36Z","lastTransitionTime":"2025-10-02T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.174522 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.174570 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.174582 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.174606 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.174619 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:36Z","lastTransitionTime":"2025-10-02T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.277977 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.278041 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.278054 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.278079 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.278095 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:36Z","lastTransitionTime":"2025-10-02T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.380890 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.380937 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.380948 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.380968 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.381013 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:36Z","lastTransitionTime":"2025-10-02T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.472507 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:36 crc kubenswrapper[4771]: E1002 09:37:36.472733 4771 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:37:36 crc kubenswrapper[4771]: E1002 09:37:36.472868 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs podName:20d5431d-4595-4b96-a8b8-8953e3dffb53 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:40.472840494 +0000 UTC m=+48.120525561 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs") pod "network-metrics-daemon-zp7kh" (UID: "20d5431d-4595-4b96-a8b8-8953e3dffb53") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.483255 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.483288 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.483298 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.483315 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.483327 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:36Z","lastTransitionTime":"2025-10-02T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.586862 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.586939 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.586955 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.586973 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.586985 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:36Z","lastTransitionTime":"2025-10-02T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.680281 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.680361 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:36 crc kubenswrapper[4771]: E1002 09:37:36.680466 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.680479 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:36 crc kubenswrapper[4771]: E1002 09:37:36.680580 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.680768 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:36 crc kubenswrapper[4771]: E1002 09:37:36.680798 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:36 crc kubenswrapper[4771]: E1002 09:37:36.681050 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.688923 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.689141 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.689240 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.689313 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.689378 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:36Z","lastTransitionTime":"2025-10-02T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.792085 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.792456 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.792518 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.792627 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.792720 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:36Z","lastTransitionTime":"2025-10-02T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.895821 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.896198 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.896300 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.896435 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.896531 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:36Z","lastTransitionTime":"2025-10-02T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.900053 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.999804 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.999896 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:36 crc kubenswrapper[4771]: I1002 09:37:36.999914 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:36.999941 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:36.999959 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:36Z","lastTransitionTime":"2025-10-02T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.103289 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.103351 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.103381 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.103411 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.103434 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:37Z","lastTransitionTime":"2025-10-02T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.206288 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.206340 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.206351 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.206371 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.206384 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:37Z","lastTransitionTime":"2025-10-02T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.309569 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.309633 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.309648 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.309671 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.309690 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:37Z","lastTransitionTime":"2025-10-02T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.413279 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.413329 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.413338 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.413357 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.413376 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:37Z","lastTransitionTime":"2025-10-02T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.516587 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.516639 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.516649 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.516707 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.516726 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:37Z","lastTransitionTime":"2025-10-02T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.619871 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.619921 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.619935 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.619954 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.619973 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:37Z","lastTransitionTime":"2025-10-02T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.723205 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.723265 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.723276 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.723296 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.723309 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:37Z","lastTransitionTime":"2025-10-02T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.826545 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.826597 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.826608 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.826626 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.826638 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:37Z","lastTransitionTime":"2025-10-02T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.929079 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.929142 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.929153 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.929169 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:37 crc kubenswrapper[4771]: I1002 09:37:37.929181 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:37Z","lastTransitionTime":"2025-10-02T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.032252 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.032315 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.032329 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.032348 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.032365 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:38Z","lastTransitionTime":"2025-10-02T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.136346 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.136409 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.136422 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.136444 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.136459 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:38Z","lastTransitionTime":"2025-10-02T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.239044 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.239099 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.239108 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.239146 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.239157 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:38Z","lastTransitionTime":"2025-10-02T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.342759 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.342823 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.342838 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.342866 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.342880 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:38Z","lastTransitionTime":"2025-10-02T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.446160 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.446211 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.446223 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.446242 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.446252 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:38Z","lastTransitionTime":"2025-10-02T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.549973 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.550037 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.550054 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.550085 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.550105 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:38Z","lastTransitionTime":"2025-10-02T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.652822 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.652878 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.652893 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.652913 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.652928 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:38Z","lastTransitionTime":"2025-10-02T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.680358 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.680420 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.680454 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.680528 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:38 crc kubenswrapper[4771]: E1002 09:37:38.680746 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:38 crc kubenswrapper[4771]: E1002 09:37:38.680848 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:38 crc kubenswrapper[4771]: E1002 09:37:38.680936 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:38 crc kubenswrapper[4771]: E1002 09:37:38.681029 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.756794 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.756841 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.756855 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.756880 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.756898 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:38Z","lastTransitionTime":"2025-10-02T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.859554 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.859606 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.859617 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.859633 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.859644 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:38Z","lastTransitionTime":"2025-10-02T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.962420 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.962476 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.962487 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.962506 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:38 crc kubenswrapper[4771]: I1002 09:37:38.962521 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:38Z","lastTransitionTime":"2025-10-02T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.067353 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.067405 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.067418 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.067438 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.067452 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:39Z","lastTransitionTime":"2025-10-02T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.170846 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.170889 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.170901 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.170919 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.170934 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:39Z","lastTransitionTime":"2025-10-02T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.273160 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.273195 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.273204 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.273238 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.273249 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:39Z","lastTransitionTime":"2025-10-02T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.375635 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.375682 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.375694 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.375714 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.375727 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:39Z","lastTransitionTime":"2025-10-02T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.478628 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.478715 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.478734 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.478795 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.478815 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:39Z","lastTransitionTime":"2025-10-02T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.581727 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.581786 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.581802 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.581825 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.581839 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:39Z","lastTransitionTime":"2025-10-02T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.685571 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.685643 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.685664 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.685685 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.685703 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:39Z","lastTransitionTime":"2025-10-02T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.789315 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.789365 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.789375 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.789396 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.789408 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:39Z","lastTransitionTime":"2025-10-02T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.892604 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.892661 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.892674 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.892698 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.892711 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:39Z","lastTransitionTime":"2025-10-02T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.996275 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.996314 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.996324 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.996343 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:39 crc kubenswrapper[4771]: I1002 09:37:39.996357 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:39Z","lastTransitionTime":"2025-10-02T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.099572 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.099643 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.099657 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.099682 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.099702 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:40Z","lastTransitionTime":"2025-10-02T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.202578 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.202629 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.202641 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.202662 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.202678 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:40Z","lastTransitionTime":"2025-10-02T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.307857 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.307892 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.307900 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.307916 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.307926 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:40Z","lastTransitionTime":"2025-10-02T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.411073 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.411189 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.411207 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.411234 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.411251 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:40Z","lastTransitionTime":"2025-10-02T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.518570 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:40 crc kubenswrapper[4771]: E1002 09:37:40.518748 4771 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:37:40 crc kubenswrapper[4771]: E1002 09:37:40.518824 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs podName:20d5431d-4595-4b96-a8b8-8953e3dffb53 nodeName:}" failed. No retries permitted until 2025-10-02 09:37:48.518798092 +0000 UTC m=+56.166483179 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs") pod "network-metrics-daemon-zp7kh" (UID: "20d5431d-4595-4b96-a8b8-8953e3dffb53") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.519639 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.519697 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.519707 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.519726 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.519741 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:40Z","lastTransitionTime":"2025-10-02T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.622756 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.622825 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.622838 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.622863 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.622879 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:40Z","lastTransitionTime":"2025-10-02T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.681478 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.681631 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:40 crc kubenswrapper[4771]: E1002 09:37:40.681711 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.681508 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.681514 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:40 crc kubenswrapper[4771]: E1002 09:37:40.681903 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:40 crc kubenswrapper[4771]: E1002 09:37:40.682022 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:40 crc kubenswrapper[4771]: E1002 09:37:40.682211 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.726521 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.726581 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.726597 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.726621 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.726636 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:40Z","lastTransitionTime":"2025-10-02T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.830003 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.830076 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.830095 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.830114 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.830162 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:40Z","lastTransitionTime":"2025-10-02T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.933281 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.933363 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.933401 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.933426 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:40 crc kubenswrapper[4771]: I1002 09:37:40.933442 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:40Z","lastTransitionTime":"2025-10-02T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.037292 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.037347 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.037361 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.037387 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.037403 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.141101 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.141179 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.141188 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.141206 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.141218 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.160079 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.160145 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.160157 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.160171 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.160181 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: E1002 09:37:41.174305 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.179442 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.179473 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.179485 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.179502 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.179515 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: E1002 09:37:41.195924 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.200436 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.200474 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.200488 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.200507 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.200523 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: E1002 09:37:41.221049 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.226705 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.226789 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.226813 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.226845 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.226866 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: E1002 09:37:41.242956 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.247560 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.247625 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.247642 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.247668 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.247687 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: E1002 09:37:41.265392 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:41 crc kubenswrapper[4771]: E1002 09:37:41.265622 4771 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.267876 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.267934 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.267951 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.267978 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.267997 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.371358 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.371409 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.371422 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.371439 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.371451 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.473831 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.473896 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.473921 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.473969 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.473998 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.580551 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.580635 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.580649 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.580676 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.580700 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.683026 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.683067 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.683079 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.683096 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.683108 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.785996 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.786042 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.786053 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.786072 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.786087 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.889154 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.889198 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.889210 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.889227 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.889239 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.992222 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.992260 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.992272 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.992289 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:41 crc kubenswrapper[4771]: I1002 09:37:41.992300 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:41Z","lastTransitionTime":"2025-10-02T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.094938 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.094974 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.094984 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.094997 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.095007 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:42Z","lastTransitionTime":"2025-10-02T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.198340 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.198411 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.198424 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.198450 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.198464 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:42Z","lastTransitionTime":"2025-10-02T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.300636 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.300680 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.300689 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.300705 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.300716 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:42Z","lastTransitionTime":"2025-10-02T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.402674 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.402725 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.402734 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.402749 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.402761 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:42Z","lastTransitionTime":"2025-10-02T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.505567 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.505603 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.505611 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.505626 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.505635 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:42Z","lastTransitionTime":"2025-10-02T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.608289 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.608329 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.608342 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.608363 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.608376 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:42Z","lastTransitionTime":"2025-10-02T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.680590 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.680588 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.680610 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.680815 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:42 crc kubenswrapper[4771]: E1002 09:37:42.680927 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:42 crc kubenswrapper[4771]: E1002 09:37:42.681086 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:42 crc kubenswrapper[4771]: E1002 09:37:42.681113 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:42 crc kubenswrapper[4771]: E1002 09:37:42.681330 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.710988 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.711056 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.711070 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.711091 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.711104 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:42Z","lastTransitionTime":"2025-10-02T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.814268 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.814324 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.814337 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.814358 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.814371 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:42Z","lastTransitionTime":"2025-10-02T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.917523 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.917596 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.917614 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.917649 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:42 crc kubenswrapper[4771]: I1002 09:37:42.917673 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:42Z","lastTransitionTime":"2025-10-02T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.019978 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.020014 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.020024 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.020039 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.020048 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:43Z","lastTransitionTime":"2025-10-02T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.122688 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.122723 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.122734 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.122748 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.122758 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:43Z","lastTransitionTime":"2025-10-02T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.225924 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.225993 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.226006 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.226027 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.226039 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:43Z","lastTransitionTime":"2025-10-02T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.328863 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.328919 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.328938 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.328963 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.328980 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:43Z","lastTransitionTime":"2025-10-02T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.432499 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.432544 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.432558 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.432575 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.432586 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:43Z","lastTransitionTime":"2025-10-02T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.535995 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.536076 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.536103 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.536170 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.536193 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:43Z","lastTransitionTime":"2025-10-02T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.640003 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.640065 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.640077 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.640098 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.640110 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:43Z","lastTransitionTime":"2025-10-02T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.699614 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.712279 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.725529 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.740088 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.743108 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.743204 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.743215 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.743233 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.743243 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:43Z","lastTransitionTime":"2025-10-02T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.753853 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.771892 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.786066 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.807985 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.823972 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.842085 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.846102 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.846165 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.846230 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.846252 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.846266 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:43Z","lastTransitionTime":"2025-10-02T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.857513 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.872784 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.895192 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.917273 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.931591 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.949852 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.949896 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.949904 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.949920 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.949929 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:43Z","lastTransitionTime":"2025-10-02T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.952198 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://716a5799b2d82f9ed214fb7e99664d287d1f8442d1af3cd4f41bf57b605000d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:28Z\\\",\\\"message\\\":\\\"bernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:37:28.973306 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:37:28.973599 6111 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:37:28.974698 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 09:37:28.974814 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 09:37:28.974871 6111 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 09:37:28.974901 6111 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 09:37:28.974955 6111 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 09:37:28.974898 6111 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 09:37:28.974881 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 09:37:28.975049 6111 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 09:37:28.975084 6111 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 09:37:28.975417 6111 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 09:37:28.975587 6111 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"points{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:37:31.669786 6287 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669793 6287 services_controller.go:445] Built service openshift-marketplace/redhat-marketplace LB template configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669805 6287 services_controller.go:451] Built service openshift-marketplace/redhat-marketplace cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-marketplace_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 09:37:31.669822 6287 services_controller.go:452] Built service openshift-marketplace/redhat-marketplace per-node LB for\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:43 crc kubenswrapper[4771]: I1002 09:37:43.966423 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:43Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.051808 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.051849 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.051862 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.051880 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.051892 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:44Z","lastTransitionTime":"2025-10-02T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.154436 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.154495 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.154513 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.154536 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.154554 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:44Z","lastTransitionTime":"2025-10-02T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.230350 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.231329 4771 scope.go:117] "RemoveContainer" containerID="208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.248602 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.257277 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.257831 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.257849 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.257873 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.257889 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:44Z","lastTransitionTime":"2025-10-02T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.266946 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.284033 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.301424 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.316475 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.331337 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.344330 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.361060 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.361218 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.361242 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.361617 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.361881 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:44Z","lastTransitionTime":"2025-10-02T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.361320 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.380814 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.402320 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.417416 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.431777 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.451869 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.465597 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.465663 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.465684 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.465710 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.465731 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:44Z","lastTransitionTime":"2025-10-02T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.466585 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.481922 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.502713 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"points{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:37:31.669786 6287 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669793 6287 services_controller.go:445] Built service openshift-marketplace/redhat-marketplace LB template configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669805 6287 services_controller.go:451] Built service openshift-marketplace/redhat-marketplace cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-marketplace_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 09:37:31.669822 6287 services_controller.go:452] Built service openshift-marketplace/redhat-marketplace per-node LB for\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.517851 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:44Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.568207 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.568277 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.568303 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.568338 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.568365 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:44Z","lastTransitionTime":"2025-10-02T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.671750 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.671782 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.671791 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.671806 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.671835 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:44Z","lastTransitionTime":"2025-10-02T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.680660 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:44 crc kubenswrapper[4771]: E1002 09:37:44.680824 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.681262 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:44 crc kubenswrapper[4771]: E1002 09:37:44.681326 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.681365 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:44 crc kubenswrapper[4771]: E1002 09:37:44.681410 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.681449 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:44 crc kubenswrapper[4771]: E1002 09:37:44.681488 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.774205 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.774243 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.774253 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.774271 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.774284 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:44Z","lastTransitionTime":"2025-10-02T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.876561 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.876619 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.876636 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.876656 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.876672 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:44Z","lastTransitionTime":"2025-10-02T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.979325 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.979383 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.979398 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.979417 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:44 crc kubenswrapper[4771]: I1002 09:37:44.979431 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:44Z","lastTransitionTime":"2025-10-02T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.067404 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/1.log" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.070727 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerStarted","Data":"813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8"} Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.071331 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.081566 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.081601 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.081614 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.081628 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.081640 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:45Z","lastTransitionTime":"2025-10-02T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.084543 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.098974 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.116030 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.133442 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.152374 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.175889 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.185967 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.186025 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.186036 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.186057 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.186068 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:45Z","lastTransitionTime":"2025-10-02T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.204275 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.222555 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.248281 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"points{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:37:31.669786 6287 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669793 6287 services_controller.go:445] Built service openshift-marketplace/redhat-marketplace LB template configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669805 6287 services_controller.go:451] Built service openshift-marketplace/redhat-marketplace cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-marketplace_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 09:37:31.669822 6287 services_controller.go:452] Built service openshift-marketplace/redhat-marketplace per-node LB for\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.263590 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.276075 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.288522 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.289100 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.289152 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.289165 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.289187 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.289200 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:45Z","lastTransitionTime":"2025-10-02T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.301288 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.312248 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.323880 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.335427 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.345705 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.391903 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.391958 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.391968 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.391985 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.392000 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:45Z","lastTransitionTime":"2025-10-02T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.495307 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.495350 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.495362 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.495381 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.495395 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:45Z","lastTransitionTime":"2025-10-02T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.598585 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.598638 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.598651 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.598670 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.598683 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:45Z","lastTransitionTime":"2025-10-02T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.701571 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.701641 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.701662 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.701693 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.701715 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:45Z","lastTransitionTime":"2025-10-02T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.804902 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.804948 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.804959 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.804975 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.804987 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:45Z","lastTransitionTime":"2025-10-02T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.907769 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.907809 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.907820 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.907836 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:45 crc kubenswrapper[4771]: I1002 09:37:45.907847 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:45Z","lastTransitionTime":"2025-10-02T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.011347 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.011396 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.011408 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.011424 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.011437 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:46Z","lastTransitionTime":"2025-10-02T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.076387 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/2.log" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.076997 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/1.log" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.080758 4771 generic.go:334] "Generic (PLEG): container finished" podID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerID="813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8" exitCode=1 Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.080825 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8"} Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.080886 4771 scope.go:117] "RemoveContainer" containerID="208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.082199 4771 scope.go:117] "RemoveContainer" containerID="813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8" Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.082554 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.102318 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.115334 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.115393 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.115407 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.115428 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.115440 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:46Z","lastTransitionTime":"2025-10-02T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.136356 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.168958 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.187053 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.202770 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.217054 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.217887 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.217915 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.217926 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.217941 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.217950 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:46Z","lastTransitionTime":"2025-10-02T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.232877 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.254301 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.273606 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.291285 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.305683 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.320204 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.320267 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.320281 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.320302 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.320317 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:46Z","lastTransitionTime":"2025-10-02T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.326075 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.344519 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.372688 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"points{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:37:31.669786 6287 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669793 6287 services_controller.go:445] Built service openshift-marketplace/redhat-marketplace LB template configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669805 6287 services_controller.go:451] Built service openshift-marketplace/redhat-marketplace cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-marketplace_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 09:37:31.669822 6287 services_controller.go:452] Built service openshift-marketplace/redhat-marketplace per-node LB for\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:45Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 09:37:45.168546 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.385466 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.385906 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:38:18.385865319 +0000 UTC m=+86.033550426 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.385993 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.386043 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.386071 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.386251 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.386272 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.386284 4771 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.386324 4771 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.386330 4771 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.386331 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:38:18.386318181 +0000 UTC m=+86.034003248 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.386496 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:38:18.386438455 +0000 UTC m=+86.034123522 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.386526 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:38:18.386512557 +0000 UTC m=+86.034197624 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.391219 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.401444 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.414104 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.423544 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.423623 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.423641 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.423661 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.423673 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:46Z","lastTransitionTime":"2025-10-02T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.487170 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.487370 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.487395 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.487408 4771 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.487462 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:38:18.487448786 +0000 UTC m=+86.135133853 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.526695 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.526737 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.526747 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.526761 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.526773 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:46Z","lastTransitionTime":"2025-10-02T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.629451 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.629490 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.629500 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.629514 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.629525 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:46Z","lastTransitionTime":"2025-10-02T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.680524 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.680524 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.680549 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.680563 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.680703 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.680740 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.680807 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:46 crc kubenswrapper[4771]: E1002 09:37:46.680861 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.732579 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.732652 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.732668 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.732689 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.732705 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:46Z","lastTransitionTime":"2025-10-02T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.834971 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.835014 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.835029 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.835047 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.835059 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:46Z","lastTransitionTime":"2025-10-02T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.905043 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.921680 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.937350 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.938180 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.938207 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.938216 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.938232 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.938250 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:46Z","lastTransitionTime":"2025-10-02T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.954293 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.970728 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:46 crc kubenswrapper[4771]: I1002 09:37:46.992776 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.010053 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.022817 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.038096 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.040981 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.041040 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.041051 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.041067 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.041076 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:47Z","lastTransitionTime":"2025-10-02T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.075286 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://208b31af4f9f550dae8afe8e14753d5231e1167b42c22335fcfb9a120ad37eee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"message\\\":\\\"points{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:37:31.669786 6287 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669793 6287 services_controller.go:445] Built service openshift-marketplace/redhat-marketplace LB template configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:37:31.669805 6287 services_controller.go:451] Built service openshift-marketplace/redhat-marketplace cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-marketplace_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1002 09:37:31.669822 6287 services_controller.go:452] Built service openshift-marketplace/redhat-marketplace per-node LB for\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:45Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 09:37:45.168546 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.086415 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/2.log" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.091873 4771 scope.go:117] "RemoveContainer" containerID="813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8" Oct 02 09:37:47 crc kubenswrapper[4771]: E1002 09:37:47.092175 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.099758 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.115148 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.128329 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.143387 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.143429 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.143443 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.143462 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.143475 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:47Z","lastTransitionTime":"2025-10-02T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.143634 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.156808 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.170773 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.182854 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.197273 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.209286 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.221352 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.235253 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.245709 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.245750 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.245759 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.245775 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.245786 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:47Z","lastTransitionTime":"2025-10-02T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.251399 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.264231 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.273676 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.283096 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.296435 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.311269 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.322336 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.336456 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.348343 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.348378 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.348390 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.348406 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.348420 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:47Z","lastTransitionTime":"2025-10-02T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.349802 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.376023 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.391652 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.403945 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.420837 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:45Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 09:37:45.168546 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.430995 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.431558 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.438659 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.445480 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.451160 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.451199 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.451208 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.451222 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.451232 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:47Z","lastTransitionTime":"2025-10-02T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.458006 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.471222 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.480259 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.492161 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.506440 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.520675 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.546268 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.553397 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.553474 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.553489 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.553507 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.553520 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:47Z","lastTransitionTime":"2025-10-02T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.567572 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.588027 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.608639 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.625765 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.648772 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.655819 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.655871 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.655885 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.655906 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.655924 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:47Z","lastTransitionTime":"2025-10-02T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.682549 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:45Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 09:37:45.168546 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.698846 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.711832 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.723753 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.758774 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.758808 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.758822 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.758846 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.758864 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:47Z","lastTransitionTime":"2025-10-02T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.862068 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.862164 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.862185 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.862214 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.862234 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:47Z","lastTransitionTime":"2025-10-02T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.965277 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.965324 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.965335 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.965353 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:47 crc kubenswrapper[4771]: I1002 09:37:47.965381 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:47Z","lastTransitionTime":"2025-10-02T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.072095 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.072186 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.072206 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.072313 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.072370 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:48Z","lastTransitionTime":"2025-10-02T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.175383 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.175437 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.175454 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.175480 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.175501 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:48Z","lastTransitionTime":"2025-10-02T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.278648 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.278685 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.278695 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.278711 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.278723 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:48Z","lastTransitionTime":"2025-10-02T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.382412 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.382464 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.382474 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.382494 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.382504 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:48Z","lastTransitionTime":"2025-10-02T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.485860 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.485917 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.485932 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.485954 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.486695 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:48Z","lastTransitionTime":"2025-10-02T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.589436 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.589519 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.589541 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.590001 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.590031 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:48Z","lastTransitionTime":"2025-10-02T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.609305 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:48 crc kubenswrapper[4771]: E1002 09:37:48.609516 4771 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:37:48 crc kubenswrapper[4771]: E1002 09:37:48.609626 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs podName:20d5431d-4595-4b96-a8b8-8953e3dffb53 nodeName:}" failed. No retries permitted until 2025-10-02 09:38:04.609601143 +0000 UTC m=+72.257286250 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs") pod "network-metrics-daemon-zp7kh" (UID: "20d5431d-4595-4b96-a8b8-8953e3dffb53") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.680710 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:48 crc kubenswrapper[4771]: E1002 09:37:48.680919 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.681418 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.681489 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:48 crc kubenswrapper[4771]: E1002 09:37:48.681577 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:48 crc kubenswrapper[4771]: E1002 09:37:48.681673 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.681746 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:48 crc kubenswrapper[4771]: E1002 09:37:48.682113 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.693982 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.694057 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.694081 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.694114 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.694176 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:48Z","lastTransitionTime":"2025-10-02T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.797152 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.797222 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.797236 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.797257 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.797274 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:48Z","lastTransitionTime":"2025-10-02T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.899582 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.899634 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.899648 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.899668 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:48 crc kubenswrapper[4771]: I1002 09:37:48.899681 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:48Z","lastTransitionTime":"2025-10-02T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.003032 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.003772 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.004232 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.004382 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.004505 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:49Z","lastTransitionTime":"2025-10-02T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.107885 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.108294 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.108441 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.108638 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.108771 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:49Z","lastTransitionTime":"2025-10-02T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.211077 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.211438 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.211524 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.211621 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.211727 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:49Z","lastTransitionTime":"2025-10-02T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.315412 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.315454 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.315467 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.315485 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.315504 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:49Z","lastTransitionTime":"2025-10-02T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.419294 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.419340 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.419355 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.419375 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.419389 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:49Z","lastTransitionTime":"2025-10-02T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.521853 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.521898 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.521912 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.521932 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.521946 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:49Z","lastTransitionTime":"2025-10-02T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.624533 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.624577 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.624594 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.624612 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.624623 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:49Z","lastTransitionTime":"2025-10-02T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.727967 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.728026 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.728045 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.728067 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.728080 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:49Z","lastTransitionTime":"2025-10-02T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.830993 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.831030 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.831040 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.831054 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.831064 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:49Z","lastTransitionTime":"2025-10-02T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.934274 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.934786 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.934820 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.935049 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:49 crc kubenswrapper[4771]: I1002 09:37:49.935071 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:49Z","lastTransitionTime":"2025-10-02T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.038339 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.038453 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.038472 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.038498 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.038519 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:50Z","lastTransitionTime":"2025-10-02T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.142088 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.142218 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.142248 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.142276 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.142297 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:50Z","lastTransitionTime":"2025-10-02T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.245783 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.245843 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.245857 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.245876 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.245889 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:50Z","lastTransitionTime":"2025-10-02T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.349417 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.349477 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.349496 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.349524 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.349547 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:50Z","lastTransitionTime":"2025-10-02T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.452867 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.452933 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.452950 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.452981 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.453000 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:50Z","lastTransitionTime":"2025-10-02T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.557021 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.557078 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.557091 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.557114 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.557155 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:50Z","lastTransitionTime":"2025-10-02T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.660312 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.660352 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.660362 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.660379 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.660391 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:50Z","lastTransitionTime":"2025-10-02T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.681057 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.681097 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.681065 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:50 crc kubenswrapper[4771]: E1002 09:37:50.681204 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.681057 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:50 crc kubenswrapper[4771]: E1002 09:37:50.681324 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:50 crc kubenswrapper[4771]: E1002 09:37:50.681373 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:50 crc kubenswrapper[4771]: E1002 09:37:50.681604 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.763165 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.763238 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.763254 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.763278 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.763293 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:50Z","lastTransitionTime":"2025-10-02T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.865654 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.865718 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.865737 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.865760 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.865778 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:50Z","lastTransitionTime":"2025-10-02T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.968549 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.968782 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.968818 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.968856 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:50 crc kubenswrapper[4771]: I1002 09:37:50.968880 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:50Z","lastTransitionTime":"2025-10-02T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.071849 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.071906 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.071916 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.071935 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.071947 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.174578 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.174646 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.174657 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.174680 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.174697 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.277985 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.278043 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.278056 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.278078 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.278091 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.381503 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.381553 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.381563 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.381582 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.381593 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.484234 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.484281 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.484291 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.484309 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.484323 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.495391 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.495447 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.495460 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.495481 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.495493 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: E1002 09:37:51.517018 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:51Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.523413 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.523456 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.523469 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.523485 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.523500 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: E1002 09:37:51.546928 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:51Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.552795 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.552846 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.552860 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.552883 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.552899 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: E1002 09:37:51.569520 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:51Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.575353 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.575416 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.575439 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.575459 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.575470 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: E1002 09:37:51.592749 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:51Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.597236 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.597307 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.597319 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.597341 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.597354 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: E1002 09:37:51.613863 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:51Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:51 crc kubenswrapper[4771]: E1002 09:37:51.614032 4771 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.616201 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.616261 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.616277 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.616300 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.616316 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.718581 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.718657 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.718674 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.718696 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.718710 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.822095 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.822220 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.822234 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.822250 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.822262 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.925653 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.925720 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.925737 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.925764 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:51 crc kubenswrapper[4771]: I1002 09:37:51.925782 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:51Z","lastTransitionTime":"2025-10-02T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.029368 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.029431 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.029442 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.029461 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.029472 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:52Z","lastTransitionTime":"2025-10-02T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.132895 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.133306 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.133392 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.133463 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.133551 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:52Z","lastTransitionTime":"2025-10-02T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.236254 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.236282 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.236293 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.236307 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.236316 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:52Z","lastTransitionTime":"2025-10-02T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.339065 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.339160 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.339179 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.339205 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.339224 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:52Z","lastTransitionTime":"2025-10-02T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.442036 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.442078 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.442089 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.442107 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.442156 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:52Z","lastTransitionTime":"2025-10-02T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.545334 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.545643 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.545728 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.545837 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.545919 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:52Z","lastTransitionTime":"2025-10-02T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.649819 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.649874 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.649891 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.649914 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.649935 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:52Z","lastTransitionTime":"2025-10-02T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.680732 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:52 crc kubenswrapper[4771]: E1002 09:37:52.681247 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.681507 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:52 crc kubenswrapper[4771]: E1002 09:37:52.681632 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.681509 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.681512 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:52 crc kubenswrapper[4771]: E1002 09:37:52.681807 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:52 crc kubenswrapper[4771]: E1002 09:37:52.682016 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.752478 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.752522 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.752534 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.752551 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.752564 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:52Z","lastTransitionTime":"2025-10-02T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.855050 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.855100 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.855109 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.855142 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.855154 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:52Z","lastTransitionTime":"2025-10-02T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.957818 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.957865 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.957880 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.957898 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:52 crc kubenswrapper[4771]: I1002 09:37:52.957911 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:52Z","lastTransitionTime":"2025-10-02T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.062406 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.062462 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.062472 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.062491 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.062508 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:53Z","lastTransitionTime":"2025-10-02T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.166686 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.167012 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.167122 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.167263 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.167380 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:53Z","lastTransitionTime":"2025-10-02T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.270229 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.270596 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.270719 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.270836 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.270954 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:53Z","lastTransitionTime":"2025-10-02T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.374209 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.374603 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.374728 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.374855 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.374972 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:53Z","lastTransitionTime":"2025-10-02T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.478877 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.479389 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.479603 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.479769 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.479910 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:53Z","lastTransitionTime":"2025-10-02T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.583211 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.583266 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.583281 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.583300 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.583312 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:53Z","lastTransitionTime":"2025-10-02T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.687227 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.687272 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.687281 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.687297 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.687306 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:53Z","lastTransitionTime":"2025-10-02T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.694819 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.708339 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.724577 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.737437 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.753993 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.769775 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.781898 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.789852 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.789881 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.789892 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.789908 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.789924 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:53Z","lastTransitionTime":"2025-10-02T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.794372 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.805795 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.819360 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.832004 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.848268 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.869190 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.883508 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.894536 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.894860 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.895040 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.895188 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.895649 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:53Z","lastTransitionTime":"2025-10-02T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.901782 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b5f98b6-fce5-465c-a055-c0deedc31dfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6047ebfe8e452e54f1e3fde9af41485e5a527210666b8b94c395c8ceded02c42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82587f84343f1807330d05aa3632a5acdb5462cc6f11ef6cd678d32164231284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://606cc0b4bdb6eda2e3d722157debd219232d6c3c1877e4c10affe91d0bbb12a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.919862 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.942251 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:45Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 09:37:45.168546 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.956422 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:37:53Z is after 2025-08-24T17:21:41Z" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.998624 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.999001 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.999074 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.999171 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:53 crc kubenswrapper[4771]: I1002 09:37:53.999240 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:53Z","lastTransitionTime":"2025-10-02T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.103533 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.103577 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.103588 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.103611 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.103621 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:54Z","lastTransitionTime":"2025-10-02T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.206975 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.207018 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.207028 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.207047 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.207060 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:54Z","lastTransitionTime":"2025-10-02T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.309535 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.309581 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.309592 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.309607 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.309618 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:54Z","lastTransitionTime":"2025-10-02T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.412584 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.412652 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.412672 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.412697 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.412715 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:54Z","lastTransitionTime":"2025-10-02T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.515777 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.515815 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.515827 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.515846 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.515858 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:54Z","lastTransitionTime":"2025-10-02T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.618546 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.618862 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.618971 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.619082 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.619203 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:54Z","lastTransitionTime":"2025-10-02T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.680865 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.680908 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:54 crc kubenswrapper[4771]: E1002 09:37:54.681019 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.681045 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:54 crc kubenswrapper[4771]: E1002 09:37:54.681116 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.680927 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:54 crc kubenswrapper[4771]: E1002 09:37:54.681323 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:54 crc kubenswrapper[4771]: E1002 09:37:54.681449 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.722193 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.722497 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.722683 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.722808 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.722910 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:54Z","lastTransitionTime":"2025-10-02T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.826071 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.826174 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.826213 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.826247 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.826274 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:54Z","lastTransitionTime":"2025-10-02T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.929464 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.929523 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.929535 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.929553 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:54 crc kubenswrapper[4771]: I1002 09:37:54.929567 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:54Z","lastTransitionTime":"2025-10-02T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.032965 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.033217 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.033226 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.033243 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.033252 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:55Z","lastTransitionTime":"2025-10-02T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.136411 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.136461 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.136472 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.136490 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.136502 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:55Z","lastTransitionTime":"2025-10-02T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.238882 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.239265 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.239365 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.239450 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.239516 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:55Z","lastTransitionTime":"2025-10-02T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.342755 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.343152 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.343248 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.343355 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.343431 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:55Z","lastTransitionTime":"2025-10-02T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.445970 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.446011 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.446022 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.446039 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.446050 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:55Z","lastTransitionTime":"2025-10-02T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.548063 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.548108 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.548118 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.548149 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.548163 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:55Z","lastTransitionTime":"2025-10-02T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.651362 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.651414 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.651425 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.651446 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.651458 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:55Z","lastTransitionTime":"2025-10-02T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.754783 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.754831 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.754844 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.754870 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.754883 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:55Z","lastTransitionTime":"2025-10-02T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.858051 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.858119 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.858159 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.858180 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.858193 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:55Z","lastTransitionTime":"2025-10-02T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.961201 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.961242 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.961252 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.961271 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:55 crc kubenswrapper[4771]: I1002 09:37:55.961281 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:55Z","lastTransitionTime":"2025-10-02T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.063842 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.063961 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.063990 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.064032 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.064060 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:56Z","lastTransitionTime":"2025-10-02T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.167043 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.167084 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.167096 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.167113 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.167137 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:56Z","lastTransitionTime":"2025-10-02T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.269457 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.269542 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.269559 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.269609 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.269624 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:56Z","lastTransitionTime":"2025-10-02T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.372931 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.372975 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.372984 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.373004 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.373016 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:56Z","lastTransitionTime":"2025-10-02T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.475981 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.476031 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.476044 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.476062 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.476075 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:56Z","lastTransitionTime":"2025-10-02T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.579002 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.579042 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.579051 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.579066 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.579076 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:56Z","lastTransitionTime":"2025-10-02T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.680435 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.680542 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:56 crc kubenswrapper[4771]: E1002 09:37:56.680603 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:56 crc kubenswrapper[4771]: E1002 09:37:56.680938 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.681008 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:56 crc kubenswrapper[4771]: E1002 09:37:56.681199 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.681306 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:56 crc kubenswrapper[4771]: E1002 09:37:56.681382 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.681593 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.681631 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.681646 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.681675 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.681695 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:56Z","lastTransitionTime":"2025-10-02T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.784663 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.784736 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.784756 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.784788 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.784808 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:56Z","lastTransitionTime":"2025-10-02T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.888712 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.888826 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.888886 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.888919 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.888989 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:56Z","lastTransitionTime":"2025-10-02T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.992630 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.992675 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.992685 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.992704 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:56 crc kubenswrapper[4771]: I1002 09:37:56.992718 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:56Z","lastTransitionTime":"2025-10-02T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.096270 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.096321 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.096332 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.096349 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.096363 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:57Z","lastTransitionTime":"2025-10-02T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.199633 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.199680 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.199688 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.199744 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.199756 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:57Z","lastTransitionTime":"2025-10-02T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.302645 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.302690 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.302707 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.302730 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.302745 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:57Z","lastTransitionTime":"2025-10-02T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.405738 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.405828 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.405841 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.405861 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.405874 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:57Z","lastTransitionTime":"2025-10-02T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.507736 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.507773 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.507784 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.507801 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.507817 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:57Z","lastTransitionTime":"2025-10-02T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.610539 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.610588 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.610603 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.610623 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.610639 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:57Z","lastTransitionTime":"2025-10-02T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.713424 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.713462 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.713471 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.713487 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.713500 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:57Z","lastTransitionTime":"2025-10-02T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.816542 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.816605 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.816616 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.816635 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.816645 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:57Z","lastTransitionTime":"2025-10-02T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.919847 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.919911 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.919928 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.919954 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:57 crc kubenswrapper[4771]: I1002 09:37:57.919975 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:57Z","lastTransitionTime":"2025-10-02T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.022482 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.022530 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.022540 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.022559 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.022570 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:58Z","lastTransitionTime":"2025-10-02T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.125426 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.125481 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.125492 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.125513 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.125525 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:58Z","lastTransitionTime":"2025-10-02T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.228403 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.228466 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.228480 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.228506 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.228524 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:58Z","lastTransitionTime":"2025-10-02T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.331612 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.331665 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.331676 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.331721 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.331733 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:58Z","lastTransitionTime":"2025-10-02T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.434368 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.434410 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.434421 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.434437 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.434447 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:58Z","lastTransitionTime":"2025-10-02T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.538177 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.538242 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.538256 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.538283 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.538299 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:58Z","lastTransitionTime":"2025-10-02T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.641630 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.642028 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.642116 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.642213 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.642285 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:58Z","lastTransitionTime":"2025-10-02T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.680525 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.680595 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.680646 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.680615 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:37:58 crc kubenswrapper[4771]: E1002 09:37:58.680744 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:37:58 crc kubenswrapper[4771]: E1002 09:37:58.680867 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:37:58 crc kubenswrapper[4771]: E1002 09:37:58.680990 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:37:58 crc kubenswrapper[4771]: E1002 09:37:58.681118 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.744997 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.745030 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.745039 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.745054 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.745064 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:58Z","lastTransitionTime":"2025-10-02T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.847987 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.848027 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.848037 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.848055 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.848068 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:58Z","lastTransitionTime":"2025-10-02T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.950585 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.950631 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.950643 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.950662 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:58 crc kubenswrapper[4771]: I1002 09:37:58.950673 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:58Z","lastTransitionTime":"2025-10-02T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.053295 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.053346 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.053356 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.053375 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.053387 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:59Z","lastTransitionTime":"2025-10-02T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.155419 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.155480 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.155490 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.155508 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.155520 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:59Z","lastTransitionTime":"2025-10-02T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.258468 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.258508 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.258520 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.258538 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.258553 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:59Z","lastTransitionTime":"2025-10-02T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.362363 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.362430 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.362445 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.362470 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.362489 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:59Z","lastTransitionTime":"2025-10-02T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.466164 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.466806 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.466825 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.466848 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.466862 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:59Z","lastTransitionTime":"2025-10-02T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.570580 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.570657 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.570672 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.570698 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.570717 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:59Z","lastTransitionTime":"2025-10-02T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.674258 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.674307 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.674316 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.674334 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.674344 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:59Z","lastTransitionTime":"2025-10-02T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.684056 4771 scope.go:117] "RemoveContainer" containerID="813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8" Oct 02 09:37:59 crc kubenswrapper[4771]: E1002 09:37:59.684458 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.777804 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.777860 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.777874 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.777898 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.777912 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:59Z","lastTransitionTime":"2025-10-02T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.880981 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.881029 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.881041 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.881064 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.881080 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:59Z","lastTransitionTime":"2025-10-02T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.984185 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.984218 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.984229 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.984247 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:37:59 crc kubenswrapper[4771]: I1002 09:37:59.984258 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:37:59Z","lastTransitionTime":"2025-10-02T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.086420 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.086491 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.086505 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.086529 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.086543 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:00Z","lastTransitionTime":"2025-10-02T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.189754 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.189804 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.189813 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.189833 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.189850 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:00Z","lastTransitionTime":"2025-10-02T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.292586 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.292638 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.292649 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.292665 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.292675 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:00Z","lastTransitionTime":"2025-10-02T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.395735 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.395778 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.395788 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.395803 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.395814 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:00Z","lastTransitionTime":"2025-10-02T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.498896 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.498936 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.498945 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.498965 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.498976 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:00Z","lastTransitionTime":"2025-10-02T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.602182 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.602468 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.602481 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.602505 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.602520 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:00Z","lastTransitionTime":"2025-10-02T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.680664 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:00 crc kubenswrapper[4771]: E1002 09:38:00.680900 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.680970 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.681005 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.681087 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:00 crc kubenswrapper[4771]: E1002 09:38:00.681143 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:00 crc kubenswrapper[4771]: E1002 09:38:00.681174 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:00 crc kubenswrapper[4771]: E1002 09:38:00.681474 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.706261 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.706324 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.706337 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.706360 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.706374 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:00Z","lastTransitionTime":"2025-10-02T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.809344 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.809715 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.809828 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.809912 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.810095 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:00Z","lastTransitionTime":"2025-10-02T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.913088 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.913501 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.913601 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.913718 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:00 crc kubenswrapper[4771]: I1002 09:38:00.913811 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:00Z","lastTransitionTime":"2025-10-02T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.016483 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.016531 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.016542 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.016560 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.016573 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.119514 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.120345 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.120460 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.120573 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.120668 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.223051 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.223096 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.223105 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.223122 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.223147 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.326346 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.326394 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.326406 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.326422 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.326432 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.429644 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.429703 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.429720 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.429746 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.429765 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.532072 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.532140 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.532151 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.532169 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.532186 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.635068 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.635117 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.635145 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.635167 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.635179 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.738358 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.738409 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.738418 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.738435 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.738445 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.820228 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.820288 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.820301 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.820324 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.820339 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: E1002 09:38:01.834264 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.838145 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.838187 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.838199 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.838222 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.838237 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: E1002 09:38:01.849525 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.853331 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.853366 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.853375 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.853393 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.853403 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: E1002 09:38:01.865734 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.869619 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.869669 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.869683 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.869701 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.869716 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: E1002 09:38:01.883920 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.888617 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.888664 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.888709 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.888732 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.888757 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:01 crc kubenswrapper[4771]: E1002 09:38:01.906264 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:01 crc kubenswrapper[4771]: E1002 09:38:01.906460 4771 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.908789 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.908829 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.908844 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.908866 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:01 crc kubenswrapper[4771]: I1002 09:38:01.908886 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:01Z","lastTransitionTime":"2025-10-02T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.011900 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.011952 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.011970 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.011991 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.012004 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:02Z","lastTransitionTime":"2025-10-02T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.114745 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.114798 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.114810 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.114829 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.114840 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:02Z","lastTransitionTime":"2025-10-02T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.217378 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.217424 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.217436 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.217457 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.217471 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:02Z","lastTransitionTime":"2025-10-02T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.321073 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.321113 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.321138 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.321159 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.321169 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:02Z","lastTransitionTime":"2025-10-02T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.423427 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.423502 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.423538 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.423559 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.423572 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:02Z","lastTransitionTime":"2025-10-02T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.526791 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.526831 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.526839 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.526858 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.526868 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:02Z","lastTransitionTime":"2025-10-02T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.630042 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.630103 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.630113 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.630154 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.630168 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:02Z","lastTransitionTime":"2025-10-02T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.681022 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.681197 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.681255 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.681351 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:02 crc kubenswrapper[4771]: E1002 09:38:02.681360 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:02 crc kubenswrapper[4771]: E1002 09:38:02.681476 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:02 crc kubenswrapper[4771]: E1002 09:38:02.681529 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:02 crc kubenswrapper[4771]: E1002 09:38:02.681569 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.733282 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.733330 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.733343 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.733363 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.733374 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:02Z","lastTransitionTime":"2025-10-02T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.836324 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.836372 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.836386 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.836409 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.836422 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:02Z","lastTransitionTime":"2025-10-02T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.939059 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.939103 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.939114 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.939157 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:02 crc kubenswrapper[4771]: I1002 09:38:02.939170 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:02Z","lastTransitionTime":"2025-10-02T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.041926 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.041992 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.042005 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.042025 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.042037 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:03Z","lastTransitionTime":"2025-10-02T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.144371 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.144432 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.144443 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.144460 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.144474 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:03Z","lastTransitionTime":"2025-10-02T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.247317 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.247358 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.247366 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.247391 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.247401 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:03Z","lastTransitionTime":"2025-10-02T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.350354 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.350403 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.350415 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.350435 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.350448 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:03Z","lastTransitionTime":"2025-10-02T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.453279 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.453336 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.453346 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.453365 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.453387 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:03Z","lastTransitionTime":"2025-10-02T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.558697 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.558765 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.558782 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.558807 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.558821 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:03Z","lastTransitionTime":"2025-10-02T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.662377 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.662431 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.662441 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.662461 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.662474 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:03Z","lastTransitionTime":"2025-10-02T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.696457 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.709257 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.721638 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.730974 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.741976 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.755566 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.765211 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.765282 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.765296 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.765317 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.765330 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:03Z","lastTransitionTime":"2025-10-02T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.769375 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.784010 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.805299 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.843516 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.857238 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b5f98b6-fce5-465c-a055-c0deedc31dfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6047ebfe8e452e54f1e3fde9af41485e5a527210666b8b94c395c8ceded02c42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82587f84343f1807330d05aa3632a5acdb5462cc6f11ef6cd678d32164231284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://606cc0b4bdb6eda2e3d722157debd219232d6c3c1877e4c10affe91d0bbb12a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.867955 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.867994 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.868007 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.868023 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.868035 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:03Z","lastTransitionTime":"2025-10-02T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.869561 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.882359 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.895409 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.917449 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:45Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 09:37:45.168546 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.928431 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.938523 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.949226 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.970608 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.970664 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.970680 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.970705 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:03 crc kubenswrapper[4771]: I1002 09:38:03.970721 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:03Z","lastTransitionTime":"2025-10-02T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.073148 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.073188 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.073201 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.073217 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.073228 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:04Z","lastTransitionTime":"2025-10-02T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.176584 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.176638 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.176651 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.176670 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.176682 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:04Z","lastTransitionTime":"2025-10-02T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.279470 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.279535 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.279545 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.279585 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.279605 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:04Z","lastTransitionTime":"2025-10-02T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.382433 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.382480 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.382493 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.382512 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.382526 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:04Z","lastTransitionTime":"2025-10-02T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.484935 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.484982 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.484991 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.485012 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.485022 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:04Z","lastTransitionTime":"2025-10-02T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.587603 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.587653 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.587666 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.587784 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.587802 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:04Z","lastTransitionTime":"2025-10-02T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.680670 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.680747 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.680750 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.680758 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:04 crc kubenswrapper[4771]: E1002 09:38:04.680871 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:04 crc kubenswrapper[4771]: E1002 09:38:04.681027 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:04 crc kubenswrapper[4771]: E1002 09:38:04.681177 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:04 crc kubenswrapper[4771]: E1002 09:38:04.681241 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.686100 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:04 crc kubenswrapper[4771]: E1002 09:38:04.686378 4771 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:38:04 crc kubenswrapper[4771]: E1002 09:38:04.686503 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs podName:20d5431d-4595-4b96-a8b8-8953e3dffb53 nodeName:}" failed. No retries permitted until 2025-10-02 09:38:36.68647411 +0000 UTC m=+104.334159177 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs") pod "network-metrics-daemon-zp7kh" (UID: "20d5431d-4595-4b96-a8b8-8953e3dffb53") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.689941 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.689988 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.689999 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.690019 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.690032 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:04Z","lastTransitionTime":"2025-10-02T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.792728 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.792789 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.792804 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.792828 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.792841 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:04Z","lastTransitionTime":"2025-10-02T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.896027 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.896070 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.896081 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.896102 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.896112 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:04Z","lastTransitionTime":"2025-10-02T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.999751 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.999800 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:04 crc kubenswrapper[4771]: I1002 09:38:04.999813 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:04.999833 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:04.999847 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:04Z","lastTransitionTime":"2025-10-02T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.102878 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.102940 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.102955 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.102976 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.102989 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:05Z","lastTransitionTime":"2025-10-02T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.206544 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.206606 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.206620 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.206646 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.206660 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:05Z","lastTransitionTime":"2025-10-02T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.308814 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.308853 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.308862 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.308881 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.308891 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:05Z","lastTransitionTime":"2025-10-02T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.411526 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.411576 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.411594 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.411614 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.411628 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:05Z","lastTransitionTime":"2025-10-02T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.514790 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.514842 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.514852 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.514876 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.514886 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:05Z","lastTransitionTime":"2025-10-02T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.618940 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.619010 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.619030 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.619053 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.619067 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:05Z","lastTransitionTime":"2025-10-02T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.722028 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.722087 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.722097 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.722115 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.722169 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:05Z","lastTransitionTime":"2025-10-02T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.825904 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.825964 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.825978 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.826001 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.826015 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:05Z","lastTransitionTime":"2025-10-02T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.928573 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.928619 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.928630 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.928647 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:05 crc kubenswrapper[4771]: I1002 09:38:05.928659 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:05Z","lastTransitionTime":"2025-10-02T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.031345 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.031421 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.031433 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.031456 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.031469 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:06Z","lastTransitionTime":"2025-10-02T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.134074 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.134141 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.134155 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.134175 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.134190 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:06Z","lastTransitionTime":"2025-10-02T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.159742 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xh7xv_bb6b65c1-83b7-4b23-abe8-c4255c6e59f8/kube-multus/0.log" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.159797 4771 generic.go:334] "Generic (PLEG): container finished" podID="bb6b65c1-83b7-4b23-abe8-c4255c6e59f8" containerID="a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c" exitCode=1 Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.159836 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xh7xv" event={"ID":"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8","Type":"ContainerDied","Data":"a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c"} Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.160352 4771 scope.go:117] "RemoveContainer" containerID="a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.174033 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.188589 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.202973 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.218019 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.230791 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.237043 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.237083 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.237092 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.237110 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.237135 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:06Z","lastTransitionTime":"2025-10-02T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.248294 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.272453 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.288566 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.303502 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.321193 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:38:05Z\\\",\\\"message\\\":\\\"2025-10-02T09:37:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1\\\\n2025-10-02T09:37:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1 to /host/opt/cni/bin/\\\\n2025-10-02T09:37:20Z [verbose] multus-daemon started\\\\n2025-10-02T09:37:20Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:38:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.336885 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.344725 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.344761 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.344769 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.344786 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.344796 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:06Z","lastTransitionTime":"2025-10-02T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.357162 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.371910 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.384744 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b5f98b6-fce5-465c-a055-c0deedc31dfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6047ebfe8e452e54f1e3fde9af41485e5a527210666b8b94c395c8ceded02c42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82587f84343f1807330d05aa3632a5acdb5462cc6f11ef6cd678d32164231284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://606cc0b4bdb6eda2e3d722157debd219232d6c3c1877e4c10affe91d0bbb12a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.404307 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:45Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 09:37:45.168546 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.417172 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.429663 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.441289 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.447369 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.447411 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.447421 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.447438 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.447448 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:06Z","lastTransitionTime":"2025-10-02T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.550566 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.550665 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.550683 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.551102 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.551136 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:06Z","lastTransitionTime":"2025-10-02T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.654255 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.654304 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.654313 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.654331 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.654340 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:06Z","lastTransitionTime":"2025-10-02T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.680943 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.681045 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.681021 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.680936 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:06 crc kubenswrapper[4771]: E1002 09:38:06.681322 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:06 crc kubenswrapper[4771]: E1002 09:38:06.681494 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:06 crc kubenswrapper[4771]: E1002 09:38:06.681647 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:06 crc kubenswrapper[4771]: E1002 09:38:06.681917 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.757604 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.757645 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.757656 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.757671 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.757681 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:06Z","lastTransitionTime":"2025-10-02T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.860348 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.860392 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.860405 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.860423 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.860435 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:06Z","lastTransitionTime":"2025-10-02T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.963259 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.963313 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.963324 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.963342 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:06 crc kubenswrapper[4771]: I1002 09:38:06.963352 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:06Z","lastTransitionTime":"2025-10-02T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.066352 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.066407 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.066417 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.066436 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.066448 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:07Z","lastTransitionTime":"2025-10-02T09:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.165117 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xh7xv_bb6b65c1-83b7-4b23-abe8-c4255c6e59f8/kube-multus/0.log" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.165262 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xh7xv" event={"ID":"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8","Type":"ContainerStarted","Data":"f311d21ae303e154a68174a305b7f431b68a36d9a03fd66eaa55699ade89bcd9"} Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.169119 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.169190 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.169205 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.169227 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.169239 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:07Z","lastTransitionTime":"2025-10-02T09:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.181820 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.194793 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.209273 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.224193 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b5f98b6-fce5-465c-a055-c0deedc31dfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6047ebfe8e452e54f1e3fde9af41485e5a527210666b8b94c395c8ceded02c42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82587f84343f1807330d05aa3632a5acdb5462cc6f11ef6cd678d32164231284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://606cc0b4bdb6eda2e3d722157debd219232d6c3c1877e4c10affe91d0bbb12a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.239609 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.252106 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.265329 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f311d21ae303e154a68174a305b7f431b68a36d9a03fd66eaa55699ade89bcd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:38:05Z\\\",\\\"message\\\":\\\"2025-10-02T09:37:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1\\\\n2025-10-02T09:37:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1 to /host/opt/cni/bin/\\\\n2025-10-02T09:37:20Z [verbose] multus-daemon started\\\\n2025-10-02T09:37:20Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:38:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:38:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.271536 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.271968 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.272098 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.272218 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.272329 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:07Z","lastTransitionTime":"2025-10-02T09:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.281467 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.302032 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.316605 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.334928 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:45Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 09:37:45.168546 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.346514 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.363714 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.374828 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.376345 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.376391 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.376401 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.376421 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.376432 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:07Z","lastTransitionTime":"2025-10-02T09:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.387526 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.397013 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.407842 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.420188 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.479357 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.479399 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.479411 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.479432 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.479447 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:07Z","lastTransitionTime":"2025-10-02T09:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.582765 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.582826 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.582842 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.582870 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.582888 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:07Z","lastTransitionTime":"2025-10-02T09:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.685203 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.685252 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.685262 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.685279 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.685290 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:07Z","lastTransitionTime":"2025-10-02T09:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.788750 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.788808 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.788818 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.788841 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.788852 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:07Z","lastTransitionTime":"2025-10-02T09:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.891713 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.891778 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.891793 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.891816 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.891828 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:07Z","lastTransitionTime":"2025-10-02T09:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.994794 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.994856 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.994869 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.994893 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:07 crc kubenswrapper[4771]: I1002 09:38:07.994908 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:07Z","lastTransitionTime":"2025-10-02T09:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.097774 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.097827 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.097840 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.097856 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.097868 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:08Z","lastTransitionTime":"2025-10-02T09:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.201243 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.201301 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.201313 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.201334 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.201352 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:08Z","lastTransitionTime":"2025-10-02T09:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.304082 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.304153 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.304165 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.304184 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.304195 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:08Z","lastTransitionTime":"2025-10-02T09:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.406915 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.406965 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.406977 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.406997 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.407009 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:08Z","lastTransitionTime":"2025-10-02T09:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.509500 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.509547 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.509567 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.509586 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.509632 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:08Z","lastTransitionTime":"2025-10-02T09:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.612432 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.612469 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.612477 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.612492 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.612501 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:08Z","lastTransitionTime":"2025-10-02T09:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.680402 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.680396 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.680419 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:08 crc kubenswrapper[4771]: E1002 09:38:08.680698 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:08 crc kubenswrapper[4771]: E1002 09:38:08.680557 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.680426 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:08 crc kubenswrapper[4771]: E1002 09:38:08.680833 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:08 crc kubenswrapper[4771]: E1002 09:38:08.680756 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.715223 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.715292 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.715305 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.715324 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.715337 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:08Z","lastTransitionTime":"2025-10-02T09:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.818522 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.818589 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.818602 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.818625 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.818640 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:08Z","lastTransitionTime":"2025-10-02T09:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.921322 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.921375 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.921391 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.921413 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:08 crc kubenswrapper[4771]: I1002 09:38:08.921424 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:08Z","lastTransitionTime":"2025-10-02T09:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.024373 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.024409 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.024421 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.024440 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.024454 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:09Z","lastTransitionTime":"2025-10-02T09:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.127805 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.127850 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.127862 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.127881 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.127894 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:09Z","lastTransitionTime":"2025-10-02T09:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.231202 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.231252 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.231261 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.231281 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.231292 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:09Z","lastTransitionTime":"2025-10-02T09:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.334085 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.334145 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.334155 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.334172 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.334188 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:09Z","lastTransitionTime":"2025-10-02T09:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.436672 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.436719 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.436727 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.436744 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.436753 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:09Z","lastTransitionTime":"2025-10-02T09:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.540005 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.540060 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.540072 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.540097 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.540109 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:09Z","lastTransitionTime":"2025-10-02T09:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.645978 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.646016 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.646027 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.646045 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.646055 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:09Z","lastTransitionTime":"2025-10-02T09:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.750016 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.750078 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.750090 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.750108 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.750122 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:09Z","lastTransitionTime":"2025-10-02T09:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.853048 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.853096 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.853107 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.853142 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.853153 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:09Z","lastTransitionTime":"2025-10-02T09:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.956917 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.956961 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.956970 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.956985 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:09 crc kubenswrapper[4771]: I1002 09:38:09.956995 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:09Z","lastTransitionTime":"2025-10-02T09:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.060140 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.060200 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.060213 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.060233 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.060244 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:10Z","lastTransitionTime":"2025-10-02T09:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.163404 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.163452 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.163462 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.163479 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.163490 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:10Z","lastTransitionTime":"2025-10-02T09:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.266955 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.267018 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.267035 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.267066 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.267092 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:10Z","lastTransitionTime":"2025-10-02T09:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.370392 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.370445 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.370460 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.370481 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.370495 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:10Z","lastTransitionTime":"2025-10-02T09:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.473283 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.474357 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.474431 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.474519 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.474592 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:10Z","lastTransitionTime":"2025-10-02T09:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.577517 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.577554 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.577567 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.577589 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.577598 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:10Z","lastTransitionTime":"2025-10-02T09:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.679889 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.680241 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.680359 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.680431 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.680495 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:10Z","lastTransitionTime":"2025-10-02T09:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.680442 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.680448 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:10 crc kubenswrapper[4771]: E1002 09:38:10.680833 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.680539 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.680442 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:10 crc kubenswrapper[4771]: E1002 09:38:10.680907 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:10 crc kubenswrapper[4771]: E1002 09:38:10.680969 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:10 crc kubenswrapper[4771]: E1002 09:38:10.681042 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.783510 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.783564 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.783577 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.783598 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.783609 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:10Z","lastTransitionTime":"2025-10-02T09:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.886255 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.886588 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.886699 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.886830 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.886904 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:10Z","lastTransitionTime":"2025-10-02T09:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.989418 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.989780 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.989876 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.989949 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:10 crc kubenswrapper[4771]: I1002 09:38:10.990048 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:10Z","lastTransitionTime":"2025-10-02T09:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.092969 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.093367 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.093471 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.093577 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.093689 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:11Z","lastTransitionTime":"2025-10-02T09:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.195704 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.196045 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.196145 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.196234 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.196343 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:11Z","lastTransitionTime":"2025-10-02T09:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.298911 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.298962 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.298973 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.298991 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.299004 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:11Z","lastTransitionTime":"2025-10-02T09:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.401671 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.402072 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.402224 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.402320 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.402411 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:11Z","lastTransitionTime":"2025-10-02T09:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.505940 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.506025 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.506041 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.506063 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.506077 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:11Z","lastTransitionTime":"2025-10-02T09:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.608352 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.608410 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.608422 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.608439 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.608449 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:11Z","lastTransitionTime":"2025-10-02T09:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.681879 4771 scope.go:117] "RemoveContainer" containerID="813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.711713 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.711780 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.711793 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.711818 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.711833 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:11Z","lastTransitionTime":"2025-10-02T09:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.816416 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.816464 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.816477 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.816504 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.816518 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:11Z","lastTransitionTime":"2025-10-02T09:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.919567 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.919879 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.919972 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.920064 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:11 crc kubenswrapper[4771]: I1002 09:38:11.920215 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:11Z","lastTransitionTime":"2025-10-02T09:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.024338 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.024385 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.024395 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.024416 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.024429 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.127497 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.127541 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.127551 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.127567 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.127579 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.148815 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.148849 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.148857 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.148873 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.148932 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: E1002 09:38:12.162410 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.166632 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.166693 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.166703 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.166722 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.166737 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: E1002 09:38:12.178691 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.181962 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.182010 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.182023 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.182042 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.182053 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.183037 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/2.log" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.185110 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerStarted","Data":"777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.185564 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:38:12 crc kubenswrapper[4771]: E1002 09:38:12.194795 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.197843 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.199080 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.199121 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.199170 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.199185 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.199195 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.211931 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: E1002 09:38:12.214217 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.218384 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.218449 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.218462 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.218484 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.218496 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.225441 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: E1002 09:38:12.237212 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: E1002 09:38:12.237377 4771 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.239418 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.239475 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.239487 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.239509 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.239528 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.240027 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.252465 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.268951 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.287658 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.304660 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f311d21ae303e154a68174a305b7f431b68a36d9a03fd66eaa55699ade89bcd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:38:05Z\\\",\\\"message\\\":\\\"2025-10-02T09:37:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1\\\\n2025-10-02T09:37:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1 to /host/opt/cni/bin/\\\\n2025-10-02T09:37:20Z [verbose] multus-daemon started\\\\n2025-10-02T09:37:20Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:38:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:38:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.324243 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.342306 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.342351 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.342367 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.342396 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.342413 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.352686 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.370803 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.382826 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b5f98b6-fce5-465c-a055-c0deedc31dfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6047ebfe8e452e54f1e3fde9af41485e5a527210666b8b94c395c8ceded02c42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82587f84343f1807330d05aa3632a5acdb5462cc6f11ef6cd678d32164231284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://606cc0b4bdb6eda2e3d722157debd219232d6c3c1877e4c10affe91d0bbb12a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.396907 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.408062 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.428356 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:45Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 09:37:45.168546 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.443601 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.445172 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.445219 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.445232 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.445252 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.445263 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.457272 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.468767 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.552725 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.552773 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.552783 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.552800 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.552813 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.661779 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.661862 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.661877 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.662234 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.662276 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.680594 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.680614 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.680767 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:12 crc kubenswrapper[4771]: E1002 09:38:12.680773 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.680833 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:12 crc kubenswrapper[4771]: E1002 09:38:12.680979 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:12 crc kubenswrapper[4771]: E1002 09:38:12.681071 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:12 crc kubenswrapper[4771]: E1002 09:38:12.681191 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.769958 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.769999 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.770010 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.770024 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.770036 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.872610 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.872651 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.872663 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.872689 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.872705 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.975439 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.975477 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.975486 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.975501 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:12 crc kubenswrapper[4771]: I1002 09:38:12.975511 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:12Z","lastTransitionTime":"2025-10-02T09:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.078483 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.078538 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.078551 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.078572 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.078583 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:13Z","lastTransitionTime":"2025-10-02T09:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.181374 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.181448 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.181464 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.181488 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.181506 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:13Z","lastTransitionTime":"2025-10-02T09:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.190103 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/3.log" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.190812 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/2.log" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.192932 4771 generic.go:334] "Generic (PLEG): container finished" podID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerID="777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa" exitCode=1 Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.192983 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa"} Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.193037 4771 scope.go:117] "RemoveContainer" containerID="813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.193985 4771 scope.go:117] "RemoveContainer" containerID="777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa" Oct 02 09:38:13 crc kubenswrapper[4771]: E1002 09:38:13.194211 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.208642 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.221449 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b5f98b6-fce5-465c-a055-c0deedc31dfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6047ebfe8e452e54f1e3fde9af41485e5a527210666b8b94c395c8ceded02c42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82587f84343f1807330d05aa3632a5acdb5462cc6f11ef6cd678d32164231284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://606cc0b4bdb6eda2e3d722157debd219232d6c3c1877e4c10affe91d0bbb12a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.236521 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.254805 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.272372 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f311d21ae303e154a68174a305b7f431b68a36d9a03fd66eaa55699ade89bcd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:38:05Z\\\",\\\"message\\\":\\\"2025-10-02T09:37:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1\\\\n2025-10-02T09:37:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1 to /host/opt/cni/bin/\\\\n2025-10-02T09:37:20Z [verbose] multus-daemon started\\\\n2025-10-02T09:37:20Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:38:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:38:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.284048 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.284096 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.284107 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.284139 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.284151 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:13Z","lastTransitionTime":"2025-10-02T09:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.292573 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.314583 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.327540 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.345731 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:45Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 09:37:45.168546 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"ces.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.140\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:38:12.745283 6851 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nI1002 09:38:12.745318 6851 services_controller.go:444] Built service openshift-oauth-apiserver/api LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1002 09:38:12.745321 6851 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x50\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.356795 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.366879 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.378174 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.386653 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.386910 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.386984 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.387068 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.387167 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:13Z","lastTransitionTime":"2025-10-02T09:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.391238 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.402764 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.415109 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.431985 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.471168 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.489825 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.489891 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.489905 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.489924 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.489934 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:13Z","lastTransitionTime":"2025-10-02T09:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.503997 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.593322 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.594262 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.594371 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.594464 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.594557 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:13Z","lastTransitionTime":"2025-10-02T09:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.696708 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.697047 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.697088 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.697097 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.697112 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.697121 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:13Z","lastTransitionTime":"2025-10-02T09:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.711725 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.728235 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.743280 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.756432 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.777393 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.793019 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.799921 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.800343 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.800426 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.800499 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.800575 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:13Z","lastTransitionTime":"2025-10-02T09:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.808782 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.820495 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f311d21ae303e154a68174a305b7f431b68a36d9a03fd66eaa55699ade89bcd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:38:05Z\\\",\\\"message\\\":\\\"2025-10-02T09:37:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1\\\\n2025-10-02T09:37:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1 to /host/opt/cni/bin/\\\\n2025-10-02T09:37:20Z [verbose] multus-daemon started\\\\n2025-10-02T09:37:20Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:38:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:38:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.834505 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.857150 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.873395 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.887850 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b5f98b6-fce5-465c-a055-c0deedc31dfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6047ebfe8e452e54f1e3fde9af41485e5a527210666b8b94c395c8ceded02c42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82587f84343f1807330d05aa3632a5acdb5462cc6f11ef6cd678d32164231284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://606cc0b4bdb6eda2e3d722157debd219232d6c3c1877e4c10affe91d0bbb12a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.902513 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.903577 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.903616 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.903628 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.903647 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.903661 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:13Z","lastTransitionTime":"2025-10-02T09:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.922748 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://813d6389c7aa77cc05707796407effd46a7ffcfc1bd344ae4e002912e46755e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:37:45Z\\\",\\\"message\\\":\\\"\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 09:37:45.168546 6487 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"ces.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.140\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:38:12.745283 6851 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nI1002 09:38:12.745318 6851 services_controller.go:444] Built service openshift-oauth-apiserver/api LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1002 09:38:12.745321 6851 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x50\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:38:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.935672 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.949877 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:13 crc kubenswrapper[4771]: I1002 09:38:13.963976 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.006986 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.007036 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.007049 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.007067 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.007079 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:14Z","lastTransitionTime":"2025-10-02T09:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.109927 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.109975 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.109985 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.110002 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.110023 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:14Z","lastTransitionTime":"2025-10-02T09:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.198177 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/3.log" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.202015 4771 scope.go:117] "RemoveContainer" containerID="777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa" Oct 02 09:38:14 crc kubenswrapper[4771]: E1002 09:38:14.202201 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.213703 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.213771 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.213807 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.213832 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.213847 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:14Z","lastTransitionTime":"2025-10-02T09:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.217985 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.235419 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.253692 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.267211 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b5f98b6-fce5-465c-a055-c0deedc31dfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6047ebfe8e452e54f1e3fde9af41485e5a527210666b8b94c395c8ceded02c42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82587f84343f1807330d05aa3632a5acdb5462cc6f11ef6cd678d32164231284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://606cc0b4bdb6eda2e3d722157debd219232d6c3c1877e4c10affe91d0bbb12a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.282176 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.296627 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.311607 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f311d21ae303e154a68174a305b7f431b68a36d9a03fd66eaa55699ade89bcd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:38:05Z\\\",\\\"message\\\":\\\"2025-10-02T09:37:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1\\\\n2025-10-02T09:37:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1 to /host/opt/cni/bin/\\\\n2025-10-02T09:37:20Z [verbose] multus-daemon started\\\\n2025-10-02T09:37:20Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:38:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:38:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.317748 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.317823 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.317846 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.317873 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.317888 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:14Z","lastTransitionTime":"2025-10-02T09:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.331668 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.354244 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.368565 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.391245 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"ces.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.140\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:38:12.745283 6851 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nI1002 09:38:12.745318 6851 services_controller.go:444] Built service openshift-oauth-apiserver/api LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1002 09:38:12.745321 6851 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x50\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:38:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.405573 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.417909 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.420238 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.420272 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.420286 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.420318 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.420332 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:14Z","lastTransitionTime":"2025-10-02T09:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.434225 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.451806 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.467004 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.482489 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.499291 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.523309 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.523368 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.523378 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.523399 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.523411 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:14Z","lastTransitionTime":"2025-10-02T09:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.625892 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.625952 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.625968 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.625992 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.626010 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:14Z","lastTransitionTime":"2025-10-02T09:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.680784 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.680849 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.680851 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.680874 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:14 crc kubenswrapper[4771]: E1002 09:38:14.680966 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:14 crc kubenswrapper[4771]: E1002 09:38:14.681087 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:14 crc kubenswrapper[4771]: E1002 09:38:14.681266 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:14 crc kubenswrapper[4771]: E1002 09:38:14.681326 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.730651 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.730696 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.730707 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.730725 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.730737 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:14Z","lastTransitionTime":"2025-10-02T09:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.833976 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.834066 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.834091 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.834121 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.834172 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:14Z","lastTransitionTime":"2025-10-02T09:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.936793 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.936837 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.936863 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.936882 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:14 crc kubenswrapper[4771]: I1002 09:38:14.936892 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:14Z","lastTransitionTime":"2025-10-02T09:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.040663 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.040762 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.040777 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.040990 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.041000 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:15Z","lastTransitionTime":"2025-10-02T09:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.143698 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.143759 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.143773 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.143791 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.143801 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:15Z","lastTransitionTime":"2025-10-02T09:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.247083 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.247143 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.247154 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.247175 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.247191 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:15Z","lastTransitionTime":"2025-10-02T09:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.349254 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.349297 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.349308 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.349325 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.349335 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:15Z","lastTransitionTime":"2025-10-02T09:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.452241 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.452279 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.452290 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.452308 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.452320 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:15Z","lastTransitionTime":"2025-10-02T09:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.555685 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.555755 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.555776 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.555801 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.555814 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:15Z","lastTransitionTime":"2025-10-02T09:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.658936 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.658997 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.659036 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.659062 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.659078 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:15Z","lastTransitionTime":"2025-10-02T09:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.761844 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.761925 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.761944 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.761965 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.761977 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:15Z","lastTransitionTime":"2025-10-02T09:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.864691 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.864740 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.864751 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.864766 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.864775 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:15Z","lastTransitionTime":"2025-10-02T09:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.968532 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.968583 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.968593 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.968706 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:15 crc kubenswrapper[4771]: I1002 09:38:15.968719 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:15Z","lastTransitionTime":"2025-10-02T09:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.071770 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.071816 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.071838 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.071860 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.071876 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:16Z","lastTransitionTime":"2025-10-02T09:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.175267 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.175332 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.175347 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.175368 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.175379 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:16Z","lastTransitionTime":"2025-10-02T09:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.278922 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.278979 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.278996 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.279017 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.279030 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:16Z","lastTransitionTime":"2025-10-02T09:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.381985 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.382040 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.382053 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.382073 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.382086 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:16Z","lastTransitionTime":"2025-10-02T09:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.485009 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.485056 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.485068 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.485088 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.485100 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:16Z","lastTransitionTime":"2025-10-02T09:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.587882 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.587942 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.587955 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.587977 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.587991 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:16Z","lastTransitionTime":"2025-10-02T09:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.680870 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.680950 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.680950 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:16 crc kubenswrapper[4771]: E1002 09:38:16.681111 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.681235 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:16 crc kubenswrapper[4771]: E1002 09:38:16.681356 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:16 crc kubenswrapper[4771]: E1002 09:38:16.681463 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:16 crc kubenswrapper[4771]: E1002 09:38:16.681526 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.690894 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.690948 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.690961 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.690978 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.690988 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:16Z","lastTransitionTime":"2025-10-02T09:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.798457 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.798526 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.798539 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.798561 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.798576 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:16Z","lastTransitionTime":"2025-10-02T09:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.901629 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.901674 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.901683 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.901700 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:16 crc kubenswrapper[4771]: I1002 09:38:16.901714 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:16Z","lastTransitionTime":"2025-10-02T09:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.004679 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.004722 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.004734 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.004753 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.004765 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:17Z","lastTransitionTime":"2025-10-02T09:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.107399 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.107448 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.107465 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.107487 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.107498 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:17Z","lastTransitionTime":"2025-10-02T09:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.210614 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.210695 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.210707 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.210727 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.210738 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:17Z","lastTransitionTime":"2025-10-02T09:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.313872 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.313922 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.313932 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.313948 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.313996 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:17Z","lastTransitionTime":"2025-10-02T09:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.417157 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.417231 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.417247 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.417267 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.417277 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:17Z","lastTransitionTime":"2025-10-02T09:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.519508 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.519562 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.519572 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.519593 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.519603 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:17Z","lastTransitionTime":"2025-10-02T09:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.622763 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.622820 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.623068 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.623093 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.623163 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:17Z","lastTransitionTime":"2025-10-02T09:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.726152 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.726201 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.726210 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.726230 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.726239 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:17Z","lastTransitionTime":"2025-10-02T09:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.829038 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.829079 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.829089 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.829105 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.829116 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:17Z","lastTransitionTime":"2025-10-02T09:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.931687 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.931745 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.931759 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.931783 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:17 crc kubenswrapper[4771]: I1002 09:38:17.931796 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:17Z","lastTransitionTime":"2025-10-02T09:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.034733 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.034824 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.034838 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.034856 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.034869 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:18Z","lastTransitionTime":"2025-10-02T09:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.137432 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.137490 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.137500 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.137517 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.137529 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:18Z","lastTransitionTime":"2025-10-02T09:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.239932 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.239984 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.239996 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.240013 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.240026 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:18Z","lastTransitionTime":"2025-10-02T09:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.342813 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.342866 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.342878 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.342899 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.342912 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:18Z","lastTransitionTime":"2025-10-02T09:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.433653 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.433800 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.433885 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.433849816 +0000 UTC m=+150.081534883 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.433957 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.433971 4771 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.433996 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.434063 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.434050481 +0000 UTC m=+150.081735538 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.434179 4771 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.434317 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.434285058 +0000 UTC m=+150.081970125 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.434498 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.434519 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.434530 4771 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.434596 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.434578477 +0000 UTC m=+150.082263544 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.445278 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.445301 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.445310 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.445326 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.445336 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:18Z","lastTransitionTime":"2025-10-02T09:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.535115 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.535346 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.535385 4771 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.535399 4771 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.535480 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.535454539 +0000 UTC m=+150.183139806 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.548521 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.548565 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.548575 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.548591 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.548604 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:18Z","lastTransitionTime":"2025-10-02T09:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.651679 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.651738 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.651748 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.651766 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.651778 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:18Z","lastTransitionTime":"2025-10-02T09:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.680342 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.680573 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.680802 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.680930 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.681024 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.681146 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.681222 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:18 crc kubenswrapper[4771]: E1002 09:38:18.681283 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.692720 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.754757 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.754809 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.754820 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.754837 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.754847 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:18Z","lastTransitionTime":"2025-10-02T09:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.858013 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.858099 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.858112 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.858149 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.858161 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:18Z","lastTransitionTime":"2025-10-02T09:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.961203 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.961260 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.961271 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.961291 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:18 crc kubenswrapper[4771]: I1002 09:38:18.961304 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:18Z","lastTransitionTime":"2025-10-02T09:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.063737 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.063776 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.063784 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.063800 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.063810 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:19Z","lastTransitionTime":"2025-10-02T09:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.166232 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.166279 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.166293 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.166315 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.166325 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:19Z","lastTransitionTime":"2025-10-02T09:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.269661 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.269703 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.269714 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.269733 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.269742 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:19Z","lastTransitionTime":"2025-10-02T09:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.372787 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.373088 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.373170 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.373271 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.373346 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:19Z","lastTransitionTime":"2025-10-02T09:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.476111 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.476387 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.476474 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.476542 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.476605 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:19Z","lastTransitionTime":"2025-10-02T09:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.578931 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.579277 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.579357 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.579423 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.579484 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:19Z","lastTransitionTime":"2025-10-02T09:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.682080 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.682383 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.682473 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.682542 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.682602 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:19Z","lastTransitionTime":"2025-10-02T09:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.786991 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.787616 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.787691 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.787761 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.787823 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:19Z","lastTransitionTime":"2025-10-02T09:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.891172 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.891629 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.891829 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.892008 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.892184 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:19Z","lastTransitionTime":"2025-10-02T09:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.997572 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.997654 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.997673 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.997701 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:19 crc kubenswrapper[4771]: I1002 09:38:19.997721 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:19Z","lastTransitionTime":"2025-10-02T09:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.100109 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.100168 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.100177 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.100191 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.100203 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:20Z","lastTransitionTime":"2025-10-02T09:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.203351 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.203396 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.203408 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.203426 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.203439 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:20Z","lastTransitionTime":"2025-10-02T09:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.307034 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.307076 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.307088 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.307105 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.307119 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:20Z","lastTransitionTime":"2025-10-02T09:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.410688 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.410732 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.410741 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.410758 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.410770 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:20Z","lastTransitionTime":"2025-10-02T09:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.514192 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.514238 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.514248 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.514266 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.514278 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:20Z","lastTransitionTime":"2025-10-02T09:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.617587 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.617636 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.617652 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.617675 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.617689 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:20Z","lastTransitionTime":"2025-10-02T09:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.681012 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.681008 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.681162 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.681033 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:20 crc kubenswrapper[4771]: E1002 09:38:20.681208 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:20 crc kubenswrapper[4771]: E1002 09:38:20.681347 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:20 crc kubenswrapper[4771]: E1002 09:38:20.681409 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:20 crc kubenswrapper[4771]: E1002 09:38:20.681486 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.720692 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.720754 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.720766 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.720785 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.720798 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:20Z","lastTransitionTime":"2025-10-02T09:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.824250 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.824310 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.824324 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.824346 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.824360 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:20Z","lastTransitionTime":"2025-10-02T09:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.927902 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.927938 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.927948 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.927967 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:20 crc kubenswrapper[4771]: I1002 09:38:20.927978 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:20Z","lastTransitionTime":"2025-10-02T09:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.030369 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.030439 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.030454 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.030476 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.030495 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:21Z","lastTransitionTime":"2025-10-02T09:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.133724 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.133772 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.133785 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.133811 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.133826 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:21Z","lastTransitionTime":"2025-10-02T09:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.235618 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.235662 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.235672 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.235691 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.235702 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:21Z","lastTransitionTime":"2025-10-02T09:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.338001 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.338077 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.338093 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.338115 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.338153 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:21Z","lastTransitionTime":"2025-10-02T09:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.440904 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.440964 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.440976 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.440998 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.441012 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:21Z","lastTransitionTime":"2025-10-02T09:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.543658 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.543719 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.543735 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.543755 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.543771 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:21Z","lastTransitionTime":"2025-10-02T09:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.647382 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.647423 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.647432 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.647449 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.647460 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:21Z","lastTransitionTime":"2025-10-02T09:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.750464 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.750510 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.750521 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.750537 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.750550 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:21Z","lastTransitionTime":"2025-10-02T09:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.854602 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.854642 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.854654 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.854674 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.854692 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:21Z","lastTransitionTime":"2025-10-02T09:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.957309 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.957359 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.957368 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.957389 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:21 crc kubenswrapper[4771]: I1002 09:38:21.957401 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:21Z","lastTransitionTime":"2025-10-02T09:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.060157 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.060209 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.060231 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.060254 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.060264 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.163330 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.163411 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.163431 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.163461 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.163487 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.268334 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.268408 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.268431 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.268464 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.268489 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.373075 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.373497 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.373512 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.373533 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.373545 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.477021 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.477082 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.477094 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.477116 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.477154 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.528150 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.528218 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.528229 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.528247 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.528258 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: E1002 09:38:22.541522 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.546114 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.546210 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.546224 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.546243 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.546254 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: E1002 09:38:22.560643 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.565505 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.565557 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.565573 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.565597 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.565612 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: E1002 09:38:22.580234 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.584844 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.584898 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.584910 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.584932 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.584945 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: E1002 09:38:22.598842 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.603849 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.603891 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.603907 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.603927 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.603939 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: E1002 09:38:22.622465 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:22 crc kubenswrapper[4771]: E1002 09:38:22.622621 4771 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.624430 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.624459 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.624467 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.624482 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.624491 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.680672 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.680679 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.680783 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.680926 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:22 crc kubenswrapper[4771]: E1002 09:38:22.681119 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:22 crc kubenswrapper[4771]: E1002 09:38:22.681522 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:22 crc kubenswrapper[4771]: E1002 09:38:22.681621 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:22 crc kubenswrapper[4771]: E1002 09:38:22.681689 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.727626 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.727662 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.727671 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.727687 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.727699 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.830434 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.830478 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.830489 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.830505 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.830516 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.934407 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.934441 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.934451 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.934466 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:22 crc kubenswrapper[4771]: I1002 09:38:22.934476 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:22Z","lastTransitionTime":"2025-10-02T09:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.037757 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.037810 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.037824 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.037843 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.037858 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:23Z","lastTransitionTime":"2025-10-02T09:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.141662 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.141703 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.141712 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.141728 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.141739 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:23Z","lastTransitionTime":"2025-10-02T09:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.243882 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.243941 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.243952 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.243972 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.243987 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:23Z","lastTransitionTime":"2025-10-02T09:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.347314 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.347367 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.347379 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.347400 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.347414 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:23Z","lastTransitionTime":"2025-10-02T09:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.450514 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.450591 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.450608 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.450637 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.450655 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:23Z","lastTransitionTime":"2025-10-02T09:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.553539 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.554069 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.554280 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.554461 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.554626 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:23Z","lastTransitionTime":"2025-10-02T09:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.657566 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.657625 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.657635 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.657655 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.657678 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:23Z","lastTransitionTime":"2025-10-02T09:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.703185 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.718658 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.733939 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b5f98b6-fce5-465c-a055-c0deedc31dfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6047ebfe8e452e54f1e3fde9af41485e5a527210666b8b94c395c8ceded02c42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82587f84343f1807330d05aa3632a5acdb5462cc6f11ef6cd678d32164231284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://606cc0b4bdb6eda2e3d722157debd219232d6c3c1877e4c10affe91d0bbb12a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.749243 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ea8229e-7556-4c0c-b605-f629d3fdcfa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ac7ad8673c663a0665b8373b9b53d2d223d1ce7e74e04783b1d63c2f01a83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f4e8484386c5d2b52033e90f3e92d987f9e18ba5ea7e0605ddbfc96215ff3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f4e8484386c5d2b52033e90f3e92d987f9e18ba5ea7e0605ddbfc96215ff3c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.760146 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.760570 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.760908 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.761055 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.761203 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:23Z","lastTransitionTime":"2025-10-02T09:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.765465 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.780682 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.796372 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f311d21ae303e154a68174a305b7f431b68a36d9a03fd66eaa55699ade89bcd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:38:05Z\\\",\\\"message\\\":\\\"2025-10-02T09:37:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1\\\\n2025-10-02T09:37:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1 to /host/opt/cni/bin/\\\\n2025-10-02T09:37:20Z [verbose] multus-daemon started\\\\n2025-10-02T09:37:20Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:38:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:38:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.813211 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.839863 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.857638 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.876997 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.877040 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.877049 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.877065 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.877075 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:23Z","lastTransitionTime":"2025-10-02T09:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.889829 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"ces.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.140\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:38:12.745283 6851 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nI1002 09:38:12.745318 6851 services_controller.go:444] Built service openshift-oauth-apiserver/api LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1002 09:38:12.745321 6851 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x50\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:38:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.905300 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.923500 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.936619 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.951502 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.964622 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.979885 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.979935 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.979945 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.979973 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.979984 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:23Z","lastTransitionTime":"2025-10-02T09:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.980471 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:23 crc kubenswrapper[4771]: I1002 09:38:23.994562 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.007179 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.082609 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.083028 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.083097 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.083224 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.083292 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:24Z","lastTransitionTime":"2025-10-02T09:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.185723 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.185756 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.185765 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.185779 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.185787 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:24Z","lastTransitionTime":"2025-10-02T09:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.288979 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.289049 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.289065 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.289087 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.289103 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:24Z","lastTransitionTime":"2025-10-02T09:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.392654 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.392711 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.392726 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.392750 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.392763 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:24Z","lastTransitionTime":"2025-10-02T09:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.496053 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.496148 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.496160 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.496182 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.496194 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:24Z","lastTransitionTime":"2025-10-02T09:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.598714 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.598748 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.598758 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.598776 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.598785 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:24Z","lastTransitionTime":"2025-10-02T09:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.680815 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.680815 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.680867 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.681351 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:24 crc kubenswrapper[4771]: E1002 09:38:24.681443 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:24 crc kubenswrapper[4771]: E1002 09:38:24.681349 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:24 crc kubenswrapper[4771]: E1002 09:38:24.681715 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:24 crc kubenswrapper[4771]: E1002 09:38:24.681838 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.702421 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.702671 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.702702 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.702730 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.702751 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:24Z","lastTransitionTime":"2025-10-02T09:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.805822 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.805882 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.805899 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.805923 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.805937 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:24Z","lastTransitionTime":"2025-10-02T09:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.908634 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.908679 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.908689 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.908706 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:24 crc kubenswrapper[4771]: I1002 09:38:24.908719 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:24Z","lastTransitionTime":"2025-10-02T09:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.012338 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.012391 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.012405 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.012423 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.012437 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:25Z","lastTransitionTime":"2025-10-02T09:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.116050 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.116161 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.116182 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.116214 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.116235 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:25Z","lastTransitionTime":"2025-10-02T09:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.220166 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.220220 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.220229 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.220246 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.220258 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:25Z","lastTransitionTime":"2025-10-02T09:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.323759 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.323808 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.323819 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.323842 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.323854 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:25Z","lastTransitionTime":"2025-10-02T09:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.427773 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.428038 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.428164 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.428254 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.428409 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:25Z","lastTransitionTime":"2025-10-02T09:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.531610 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.531660 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.531674 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.531697 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.531712 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:25Z","lastTransitionTime":"2025-10-02T09:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.633560 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.633626 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.633637 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.633652 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.633663 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:25Z","lastTransitionTime":"2025-10-02T09:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.736224 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.736283 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.736301 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.736329 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.736349 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:25Z","lastTransitionTime":"2025-10-02T09:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.840341 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.840394 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.840403 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.840424 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.840441 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:25Z","lastTransitionTime":"2025-10-02T09:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.944006 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.944339 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.944403 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.944474 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:25 crc kubenswrapper[4771]: I1002 09:38:25.944555 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:25Z","lastTransitionTime":"2025-10-02T09:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.046881 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.046943 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.046953 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.046970 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.046980 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:26Z","lastTransitionTime":"2025-10-02T09:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.148675 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.148999 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.149089 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.149215 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.149300 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:26Z","lastTransitionTime":"2025-10-02T09:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.251459 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.251529 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.251544 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.251563 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.251573 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:26Z","lastTransitionTime":"2025-10-02T09:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.354067 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.354425 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.354501 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.354578 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.354674 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:26Z","lastTransitionTime":"2025-10-02T09:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.457872 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.458075 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.458096 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.458119 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.458169 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:26Z","lastTransitionTime":"2025-10-02T09:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.562020 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.562712 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.562880 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.563052 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.563261 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:26Z","lastTransitionTime":"2025-10-02T09:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.666612 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.666694 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.666711 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.666735 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.666749 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:26Z","lastTransitionTime":"2025-10-02T09:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.681303 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.681303 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:26 crc kubenswrapper[4771]: E1002 09:38:26.681530 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.681344 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:26 crc kubenswrapper[4771]: E1002 09:38:26.681593 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.681326 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:26 crc kubenswrapper[4771]: E1002 09:38:26.681673 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:26 crc kubenswrapper[4771]: E1002 09:38:26.681763 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.769232 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.769649 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.769734 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.769845 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.769933 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:26Z","lastTransitionTime":"2025-10-02T09:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.874383 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.875076 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.875745 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.875955 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.876107 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:26Z","lastTransitionTime":"2025-10-02T09:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.979503 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.979571 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.979584 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.979606 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:26 crc kubenswrapper[4771]: I1002 09:38:26.979622 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:26Z","lastTransitionTime":"2025-10-02T09:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.082843 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.082892 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.082905 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.082924 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.082937 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:27Z","lastTransitionTime":"2025-10-02T09:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.185631 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.185669 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.185681 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.185697 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.185708 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:27Z","lastTransitionTime":"2025-10-02T09:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.288007 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.288036 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.288046 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.288059 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.288069 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:27Z","lastTransitionTime":"2025-10-02T09:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.391228 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.391302 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.391319 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.391345 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.391361 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:27Z","lastTransitionTime":"2025-10-02T09:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.494360 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.494418 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.494432 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.494449 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.494465 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:27Z","lastTransitionTime":"2025-10-02T09:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.597345 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.597393 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.597404 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.597423 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.597435 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:27Z","lastTransitionTime":"2025-10-02T09:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.700159 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.700222 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.700235 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.700258 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.700275 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:27Z","lastTransitionTime":"2025-10-02T09:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.803992 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.804081 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.804095 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.804116 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.804173 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:27Z","lastTransitionTime":"2025-10-02T09:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.907247 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.907321 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.907334 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.907354 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:27 crc kubenswrapper[4771]: I1002 09:38:27.907392 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:27Z","lastTransitionTime":"2025-10-02T09:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.010697 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.010783 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.010836 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.010864 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.010883 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:28Z","lastTransitionTime":"2025-10-02T09:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.114239 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.114292 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.114309 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.114333 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.114349 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:28Z","lastTransitionTime":"2025-10-02T09:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.216886 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.216926 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.216934 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.216948 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.216958 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:28Z","lastTransitionTime":"2025-10-02T09:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.320182 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.320252 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.320270 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.320295 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.320312 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:28Z","lastTransitionTime":"2025-10-02T09:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.423488 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.423541 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.423554 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.423574 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.423590 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:28Z","lastTransitionTime":"2025-10-02T09:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.526978 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.527052 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.527071 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.527097 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.527114 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:28Z","lastTransitionTime":"2025-10-02T09:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.630054 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.630114 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.630153 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.630173 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.630185 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:28Z","lastTransitionTime":"2025-10-02T09:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.681056 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.681154 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.681085 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.681332 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:28 crc kubenswrapper[4771]: E1002 09:38:28.681527 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:28 crc kubenswrapper[4771]: E1002 09:38:28.681721 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:28 crc kubenswrapper[4771]: E1002 09:38:28.681813 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:28 crc kubenswrapper[4771]: E1002 09:38:28.681967 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.733406 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.733472 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.733491 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.733522 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.733554 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:28Z","lastTransitionTime":"2025-10-02T09:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.837336 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.837435 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.837463 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.837506 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.837534 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:28Z","lastTransitionTime":"2025-10-02T09:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.940580 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.940625 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.940637 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.940655 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:28 crc kubenswrapper[4771]: I1002 09:38:28.940667 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:28Z","lastTransitionTime":"2025-10-02T09:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.043868 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.043944 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.043955 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.043972 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.043982 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:29Z","lastTransitionTime":"2025-10-02T09:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.147114 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.147186 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.147200 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.147221 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.147235 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:29Z","lastTransitionTime":"2025-10-02T09:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.250467 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.250901 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.251256 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.251504 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.251680 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:29Z","lastTransitionTime":"2025-10-02T09:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.355305 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.355805 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.355902 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.355999 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.356079 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:29Z","lastTransitionTime":"2025-10-02T09:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.460451 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.461223 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.461361 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.461492 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.461598 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:29Z","lastTransitionTime":"2025-10-02T09:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.565551 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.565631 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.565642 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.565661 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.565675 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:29Z","lastTransitionTime":"2025-10-02T09:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.669053 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.669116 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.669161 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.669188 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.669205 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:29Z","lastTransitionTime":"2025-10-02T09:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.682343 4771 scope.go:117] "RemoveContainer" containerID="777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa" Oct 02 09:38:29 crc kubenswrapper[4771]: E1002 09:38:29.683037 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.772042 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.772106 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.772117 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.772152 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.772163 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:29Z","lastTransitionTime":"2025-10-02T09:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.876109 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.876233 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.876260 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.876297 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.876322 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:29Z","lastTransitionTime":"2025-10-02T09:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.978874 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.978933 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.978943 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.978960 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:29 crc kubenswrapper[4771]: I1002 09:38:29.978973 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:29Z","lastTransitionTime":"2025-10-02T09:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.082174 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.082265 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.082282 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.082306 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.082324 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:30Z","lastTransitionTime":"2025-10-02T09:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.185469 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.185534 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.185551 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.185578 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.185594 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:30Z","lastTransitionTime":"2025-10-02T09:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.288278 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.288740 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.288850 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.288962 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.289060 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:30Z","lastTransitionTime":"2025-10-02T09:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.392347 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.392427 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.392441 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.392468 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.392486 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:30Z","lastTransitionTime":"2025-10-02T09:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.495752 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.496205 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.496274 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.496357 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.496428 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:30Z","lastTransitionTime":"2025-10-02T09:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.599345 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.599394 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.599405 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.599423 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.599436 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:30Z","lastTransitionTime":"2025-10-02T09:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.681103 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.681223 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.681227 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:30 crc kubenswrapper[4771]: E1002 09:38:30.681882 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.681354 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:30 crc kubenswrapper[4771]: E1002 09:38:30.681956 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:30 crc kubenswrapper[4771]: E1002 09:38:30.681645 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:30 crc kubenswrapper[4771]: E1002 09:38:30.682163 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.703029 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.703155 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.703170 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.703195 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.703208 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:30Z","lastTransitionTime":"2025-10-02T09:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.806602 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.807036 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.807439 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.807695 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.807913 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:30Z","lastTransitionTime":"2025-10-02T09:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.911739 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.912056 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.912160 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.912252 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:30 crc kubenswrapper[4771]: I1002 09:38:30.912395 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:30Z","lastTransitionTime":"2025-10-02T09:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.016022 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.017328 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.017501 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.017602 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.017685 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:31Z","lastTransitionTime":"2025-10-02T09:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.120780 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.120824 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.120833 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.120850 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.120861 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:31Z","lastTransitionTime":"2025-10-02T09:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.225328 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.225381 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.225393 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.225409 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.225419 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:31Z","lastTransitionTime":"2025-10-02T09:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.328436 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.328813 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.328968 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.329068 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.329175 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:31Z","lastTransitionTime":"2025-10-02T09:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.433268 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.433332 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.433343 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.433365 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.433379 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:31Z","lastTransitionTime":"2025-10-02T09:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.536512 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.536558 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.536569 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.536587 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.536599 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:31Z","lastTransitionTime":"2025-10-02T09:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.640750 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.640808 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.640822 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.640842 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.640856 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:31Z","lastTransitionTime":"2025-10-02T09:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.744221 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.744276 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.744290 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.744310 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.744322 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:31Z","lastTransitionTime":"2025-10-02T09:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.847255 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.847687 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.847866 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.847967 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.848041 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:31Z","lastTransitionTime":"2025-10-02T09:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.953631 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.953673 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.953682 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.953701 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:31 crc kubenswrapper[4771]: I1002 09:38:31.953711 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:31Z","lastTransitionTime":"2025-10-02T09:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.056611 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.056677 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.056689 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.056710 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.056729 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:32Z","lastTransitionTime":"2025-10-02T09:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.159488 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.159532 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.159541 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.159559 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.159569 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:32Z","lastTransitionTime":"2025-10-02T09:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.261528 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.261576 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.261587 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.261612 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.261625 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:32Z","lastTransitionTime":"2025-10-02T09:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.365777 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.365873 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.365901 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.365937 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.365969 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:32Z","lastTransitionTime":"2025-10-02T09:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.469388 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.473913 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.473973 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.474001 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.474022 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:32Z","lastTransitionTime":"2025-10-02T09:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.577592 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.577688 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.577709 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.577740 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.577760 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:32Z","lastTransitionTime":"2025-10-02T09:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.680482 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:32 crc kubenswrapper[4771]: E1002 09:38:32.680674 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.680799 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:32 crc kubenswrapper[4771]: E1002 09:38:32.680871 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.680911 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.680908 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:32 crc kubenswrapper[4771]: E1002 09:38:32.681311 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:32 crc kubenswrapper[4771]: E1002 09:38:32.681433 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.681763 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.681821 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.681839 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.681872 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.681895 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:32Z","lastTransitionTime":"2025-10-02T09:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.803905 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.803954 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.803968 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.803989 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.804003 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:32Z","lastTransitionTime":"2025-10-02T09:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.906425 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.906474 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.906487 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.906506 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.906516 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:32Z","lastTransitionTime":"2025-10-02T09:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.935012 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.935066 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.935081 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.935097 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.935107 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:32Z","lastTransitionTime":"2025-10-02T09:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:32 crc kubenswrapper[4771]: E1002 09:38:32.949043 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.953618 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.953662 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.953671 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.953688 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.953697 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:32Z","lastTransitionTime":"2025-10-02T09:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:32 crc kubenswrapper[4771]: E1002 09:38:32.967226 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.971104 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.971178 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.971194 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.971221 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.971236 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:32Z","lastTransitionTime":"2025-10-02T09:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:32 crc kubenswrapper[4771]: E1002 09:38:32.990786 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.995870 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.995988 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.996056 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.996121 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:32 crc kubenswrapper[4771]: I1002 09:38:32.996207 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:32Z","lastTransitionTime":"2025-10-02T09:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:33 crc kubenswrapper[4771]: E1002 09:38:33.013086 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.018211 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.018267 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.018286 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.018318 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.018338 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:33Z","lastTransitionTime":"2025-10-02T09:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:33 crc kubenswrapper[4771]: E1002 09:38:33.032077 4771 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:38:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bea03645-5dbd-471c-a261-bbba6e916461\\\",\\\"systemUUID\\\":\\\"d79e76cf-cd2e-4589-9fea-018e1e59062c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: E1002 09:38:33.032270 4771 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.034571 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.034635 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.034652 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.034673 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.034687 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:33Z","lastTransitionTime":"2025-10-02T09:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.137739 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.137784 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.137797 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.137817 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.137830 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:33Z","lastTransitionTime":"2025-10-02T09:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.240858 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.240931 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.240944 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.240970 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.240991 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:33Z","lastTransitionTime":"2025-10-02T09:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.342837 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.342875 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.342900 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.342919 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.342931 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:33Z","lastTransitionTime":"2025-10-02T09:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.445009 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.445050 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.445059 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.445073 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.445104 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:33Z","lastTransitionTime":"2025-10-02T09:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.546875 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.546919 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.546933 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.546948 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.546957 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:33Z","lastTransitionTime":"2025-10-02T09:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.649822 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.649875 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.649886 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.649904 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.649914 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:33Z","lastTransitionTime":"2025-10-02T09:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.699361 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b09690eb-cfe8-4891-821e-1194ae265b07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384b425cc3f6642d1ec12af5d68303ea1cf28c95f96efde6dd21e9c550bca4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46f4a2757f84f23f200b1567ed2a67e73cd7b8f14b0690b0c75ad5b0e3c4f590\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1954b8c504fe9fbc3e89637742ff017ee1bd0e38785d28a866216d7a7170d875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c15e7bea3f118438df1518997600bb4a76f1d5044781f93c2c1838ef22c69a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aaae8d9e1ff74083c8ebc217b5aecfb8a6c853706017c09b919e995f75ad989d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b884f0b4d7fb5f8b4bd5428433b6b2ad58385e01000f5af690e904eada8d9a97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1a3542acee27752dc40602dd5f9bf803d3d5666d323d827300ea964436df85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mx86k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qb5kw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.721548 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44ec55a5-5f9f-4c63-a715-e2009730cc09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11770c58ff6bd6ee5aa8ec49fd7f441e7d7eeaac54e6852d597cfef82e6dd540\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f19e9033848b46232dac4fbd2c29ec2cbcc32a78f32240cc5d8d0b17c0470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a79cf485f082a73757eec3cf5b726c2691b619f643299b3d05cbaabe95fa904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc3ee68e8b8eac6fbe84f450bf7441e401a50de416c133287200b9a412aa7e3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cdc1566debe7d1345fe34695acae2979873d62727c0ca2adcb0843fb3303f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664915e194253729b0f691f74dde2f7e4f780b9026084821dbcc51323d8dfa9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4401d42e733e70f016d1a16e528bd4b58d30c4aca0bf114c1f1106c431ebde7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67141291be47667d644cefdbdb61822f1f6b1606a970d441ed4ae317e4665522\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.738149 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736229b-b3cb-4c9d-87eb-9f19e82dc74f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cd5703118bd76878c289083a6903989d1fbb6b32c3e9b3dea5a04e99078c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e08b6ad652ba5d9950ea1e70a99fc6268419cf3f78047b86dca26509c11107a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81dcaf1f9e6a5ffd9c7f140d7d613d26942ef5d073931481c6d7dd0b1f106d1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1912430d349652a44991d2c7e26f6a3f70c3009a317af21ced9f0686de860407\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c9f6e0043389fa8200f12abd5818e69d03b5f6acbec5382c310f5430cdb1355\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:37:14.193728 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:37:14.193896 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:37:14.194850 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2289376780/tls.crt::/tmp/serving-cert-2289376780/tls.key\\\\\\\"\\\\nI1002 09:37:14.478155 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:37:14.488233 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:37:14.488274 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:37:14.488330 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:37:14.488342 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:37:14.497894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:37:14.497937 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497945 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:37:14.497951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:37:14.497955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:37:14.497959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:37:14.497964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 09:37:14.498375 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 09:37:14.501247 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcb0e23385b5489d53a61789473a1f231ab2f75400189ef593900676fb3bfb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78e30f64ad130d8214d8ad17c34592a2c08920062b76d5564f76f4262ec79be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.752039 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b5f98b6-fce5-465c-a055-c0deedc31dfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6047ebfe8e452e54f1e3fde9af41485e5a527210666b8b94c395c8ceded02c42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82587f84343f1807330d05aa3632a5acdb5462cc6f11ef6cd678d32164231284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://606cc0b4bdb6eda2e3d722157debd219232d6c3c1877e4c10affe91d0bbb12a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cb0a977fa92b2856a2e30650edbd353784f31c4ab2571fb84fe28ea24b836aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.753425 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.753757 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.753942 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.754152 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.754331 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:33Z","lastTransitionTime":"2025-10-02T09:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.764327 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ea8229e-7556-4c0c-b605-f629d3fdcfa4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ac7ad8673c663a0665b8373b9b53d2d223d1ce7e74e04783b1d63c2f01a83f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f4e8484386c5d2b52033e90f3e92d987f9e18ba5ea7e0605ddbfc96215ff3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f4e8484386c5d2b52033e90f3e92d987f9e18ba5ea7e0605ddbfc96215ff3c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.781726 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.799513 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.814395 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xh7xv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:38:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f311d21ae303e154a68174a305b7f431b68a36d9a03fd66eaa55699ade89bcd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:38:05Z\\\",\\\"message\\\":\\\"2025-10-02T09:37:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1\\\\n2025-10-02T09:37:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b60f5fdb-1ddf-4ca7-b1d2-dc2916a61ad1 to /host/opt/cni/bin/\\\\n2025-10-02T09:37:20Z [verbose] multus-daemon started\\\\n2025-10-02T09:37:20Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:38:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:38:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pbs6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xh7xv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.839768 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:38:12Z\\\",\\\"message\\\":\\\"ces.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.140\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:38:12.745283 6851 services_controller.go:452] Built service openshift-kube-apiserver/apiserver per-node LB for network=default: []services.LB{}\\\\nI1002 09:38:12.745318 6851 services_controller.go:444] Built service openshift-oauth-apiserver/api LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1002 09:38:12.745321 6851 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x50\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:38:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pl2v4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4lxxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.851964 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d6fc451-a597-4cf5-99b9-8161888e8fcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a427d99ce848f22376aeebcf574e975f53496b0fc35a87f861163fe51fb0313a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ac95b4dc1bb2a85c9c1a14ff92e733d085c804123b79c95a8479f53d27a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsqb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wf5gr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.857442 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.857485 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.857497 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.857514 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.857524 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:33Z","lastTransitionTime":"2025-10-02T09:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.865480 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qfhxl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f29540bb-d982-4a75-9ee6-e24242514b05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2742d3c90063e168e9447b556bfd558d3ef112fc66f5cb5112c6ada62113113f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nw4hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qfhxl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.890389 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5431d-4595-4b96-a8b8-8953e3dffb53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bfj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zp7kh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.906917 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c64de7d-20e1-4329-a075-5fb34f52a131\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:36:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d6afdd3f911e2e974b01ccd6ed4995c4474343e1968e41dc548670e8a50c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eb41b712a28269e5da8bceff9077315ecaf88eb9a79359cc198c9f17ab77510\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfd3d2ad5f9040dfd4d9fd9b590d90f04ac758bcc1332ebf920815a825340785\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b274ff602e85efa493da27d785c024b80c8cfe74c430d91421493ffb98051e92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:36:53Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.919764 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b36d8b4977a72228ab4e9e6debdd46ebededb12414396860c4b2b39be05859c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.932372 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.940922 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5z94m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a074140a-f8fe-4e03-90e9-368c0632e369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45546fb90ffc25d9292d00cb6bb5996a726911f5bd51e0d202991d03028aa192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lhtxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5z94m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.954140 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d686b1-d29a-42c4-97ce-e239a0c680b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d3b7b7261a52418cf4e57fc0d265266163beb177f433cbf3314c3dc6797e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:37:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4mpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.963568 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.963732 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.963823 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.963923 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.964041 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:33Z","lastTransitionTime":"2025-10-02T09:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.970617 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fcae45adbb1549dc6e834b16388eaae6966132d590cdc9d8a467b845482fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4a82b906068b9efe77ef7af6246b917d9e16b992f633df6a23aede08c4e5d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:33 crc kubenswrapper[4771]: I1002 09:38:33.989607 4771 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:37:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29c2db0b8afc78b1a121720a73ef5e2f8f57a415a7a14b34444c17b6f97febd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:37:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:38:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.067307 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.067361 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.067374 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.067392 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.067405 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:34Z","lastTransitionTime":"2025-10-02T09:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.170493 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.170547 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.170566 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.170595 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.170615 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:34Z","lastTransitionTime":"2025-10-02T09:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.273914 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.274503 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.274541 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.274575 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.274600 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:34Z","lastTransitionTime":"2025-10-02T09:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.378300 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.378398 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.378418 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.378445 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.378465 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:34Z","lastTransitionTime":"2025-10-02T09:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.481376 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.481434 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.481448 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.481469 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.481483 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:34Z","lastTransitionTime":"2025-10-02T09:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.585394 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.585447 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.585458 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.585477 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.585490 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:34Z","lastTransitionTime":"2025-10-02T09:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.680568 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.680623 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.680827 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.680913 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:34 crc kubenswrapper[4771]: E1002 09:38:34.681724 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:34 crc kubenswrapper[4771]: E1002 09:38:34.681919 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:34 crc kubenswrapper[4771]: E1002 09:38:34.682052 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:34 crc kubenswrapper[4771]: E1002 09:38:34.681826 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.689216 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.689255 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.689264 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.689280 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.689292 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:34Z","lastTransitionTime":"2025-10-02T09:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.792874 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.792921 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.792932 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.792951 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.792964 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:34Z","lastTransitionTime":"2025-10-02T09:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.895336 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.895391 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.895404 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.895423 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.895440 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:34Z","lastTransitionTime":"2025-10-02T09:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.998399 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.998442 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.998455 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.998473 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:34 crc kubenswrapper[4771]: I1002 09:38:34.998486 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:34Z","lastTransitionTime":"2025-10-02T09:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.102065 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.102491 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.102582 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.102708 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.102812 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:35Z","lastTransitionTime":"2025-10-02T09:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.206272 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.206315 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.206323 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.206337 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.206347 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:35Z","lastTransitionTime":"2025-10-02T09:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.309199 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.309243 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.309284 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.309299 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.309308 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:35Z","lastTransitionTime":"2025-10-02T09:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.412214 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.412257 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.412267 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.412302 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.412314 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:35Z","lastTransitionTime":"2025-10-02T09:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.515572 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.515644 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.515655 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.515669 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.515678 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:35Z","lastTransitionTime":"2025-10-02T09:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.618558 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.618879 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.618963 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.619054 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.619119 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:35Z","lastTransitionTime":"2025-10-02T09:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.722016 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.722478 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.722550 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.722638 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.722748 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:35Z","lastTransitionTime":"2025-10-02T09:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.826092 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.826147 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.826160 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.826177 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.826190 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:35Z","lastTransitionTime":"2025-10-02T09:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.928754 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.929299 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.929408 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.929511 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:35 crc kubenswrapper[4771]: I1002 09:38:35.929598 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:35Z","lastTransitionTime":"2025-10-02T09:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.033072 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.033142 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.033153 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.033169 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.033178 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:36Z","lastTransitionTime":"2025-10-02T09:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.136153 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.136192 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.136201 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.136215 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.136224 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:36Z","lastTransitionTime":"2025-10-02T09:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.239459 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.239531 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.239548 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.239567 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.239583 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:36Z","lastTransitionTime":"2025-10-02T09:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.342761 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.342826 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.342840 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.342897 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.342915 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:36Z","lastTransitionTime":"2025-10-02T09:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.445379 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.445412 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.445420 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.445434 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.445443 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:36Z","lastTransitionTime":"2025-10-02T09:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.547759 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.547842 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.547861 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.547887 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.547903 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:36Z","lastTransitionTime":"2025-10-02T09:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.650331 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.650388 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.650400 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.650421 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.650435 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:36Z","lastTransitionTime":"2025-10-02T09:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.681092 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.681164 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.681217 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.681227 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:36 crc kubenswrapper[4771]: E1002 09:38:36.681330 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:36 crc kubenswrapper[4771]: E1002 09:38:36.681416 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:36 crc kubenswrapper[4771]: E1002 09:38:36.681586 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:36 crc kubenswrapper[4771]: E1002 09:38:36.681690 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.744571 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:36 crc kubenswrapper[4771]: E1002 09:38:36.744760 4771 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:38:36 crc kubenswrapper[4771]: E1002 09:38:36.744816 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs podName:20d5431d-4595-4b96-a8b8-8953e3dffb53 nodeName:}" failed. No retries permitted until 2025-10-02 09:39:40.744800173 +0000 UTC m=+168.392485240 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs") pod "network-metrics-daemon-zp7kh" (UID: "20d5431d-4595-4b96-a8b8-8953e3dffb53") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.752464 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.752504 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.752518 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.752535 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.752548 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:36Z","lastTransitionTime":"2025-10-02T09:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.855025 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.855101 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.855110 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.855138 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.855151 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:36Z","lastTransitionTime":"2025-10-02T09:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.957265 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.957307 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.957315 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.957331 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:36 crc kubenswrapper[4771]: I1002 09:38:36.957340 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:36Z","lastTransitionTime":"2025-10-02T09:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.058899 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.058950 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.058962 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.058977 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.058988 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:37Z","lastTransitionTime":"2025-10-02T09:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.161094 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.161164 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.161175 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.161192 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.161203 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:37Z","lastTransitionTime":"2025-10-02T09:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.265356 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.265499 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.265515 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.265540 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.265552 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:37Z","lastTransitionTime":"2025-10-02T09:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.368446 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.368485 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.368494 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.368511 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.368522 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:37Z","lastTransitionTime":"2025-10-02T09:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.470610 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.470648 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.470657 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.470672 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.470681 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:37Z","lastTransitionTime":"2025-10-02T09:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.575632 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.575683 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.575696 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.575715 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.575727 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:37Z","lastTransitionTime":"2025-10-02T09:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.678348 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.678391 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.678403 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.678417 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.678426 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:37Z","lastTransitionTime":"2025-10-02T09:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.781355 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.781414 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.781425 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.781442 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.781453 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:37Z","lastTransitionTime":"2025-10-02T09:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.884704 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.884786 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.884807 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.884837 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.884857 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:37Z","lastTransitionTime":"2025-10-02T09:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.987670 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.987720 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.987733 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.987750 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:37 crc kubenswrapper[4771]: I1002 09:38:37.987762 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:37Z","lastTransitionTime":"2025-10-02T09:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.091888 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.091973 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.091982 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.091999 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.092010 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:38Z","lastTransitionTime":"2025-10-02T09:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.194565 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.194640 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.194653 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.194674 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.194689 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:38Z","lastTransitionTime":"2025-10-02T09:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.297921 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.297973 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.297988 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.298004 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.298017 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:38Z","lastTransitionTime":"2025-10-02T09:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.400914 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.400967 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.400983 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.401000 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.401014 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:38Z","lastTransitionTime":"2025-10-02T09:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.504032 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.504079 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.504089 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.504107 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.504117 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:38Z","lastTransitionTime":"2025-10-02T09:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.607102 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.607170 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.607180 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.607251 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.607261 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:38Z","lastTransitionTime":"2025-10-02T09:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.681080 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.681097 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:38 crc kubenswrapper[4771]: E1002 09:38:38.681539 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.681208 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:38 crc kubenswrapper[4771]: E1002 09:38:38.681802 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.681148 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:38 crc kubenswrapper[4771]: E1002 09:38:38.681689 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:38 crc kubenswrapper[4771]: E1002 09:38:38.682021 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.711205 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.711270 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.711287 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.711309 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.711325 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:38Z","lastTransitionTime":"2025-10-02T09:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.814220 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.814658 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.814762 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.814895 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.815037 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:38Z","lastTransitionTime":"2025-10-02T09:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.918324 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.918401 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.918418 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.918440 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:38 crc kubenswrapper[4771]: I1002 09:38:38.918476 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:38Z","lastTransitionTime":"2025-10-02T09:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.022085 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.022153 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.022163 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.022180 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.022191 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:39Z","lastTransitionTime":"2025-10-02T09:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.124396 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.124432 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.124442 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.124457 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.124466 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:39Z","lastTransitionTime":"2025-10-02T09:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.227151 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.227549 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.227635 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.227706 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.227770 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:39Z","lastTransitionTime":"2025-10-02T09:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.330265 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.330325 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.330335 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.330364 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.330375 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:39Z","lastTransitionTime":"2025-10-02T09:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.433731 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.433802 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.433818 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.433843 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.433855 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:39Z","lastTransitionTime":"2025-10-02T09:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.537912 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.537976 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.537990 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.538011 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.538028 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:39Z","lastTransitionTime":"2025-10-02T09:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.640893 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.640941 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.640952 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.640968 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.640981 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:39Z","lastTransitionTime":"2025-10-02T09:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.743997 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.744053 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.744064 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.744081 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.744097 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:39Z","lastTransitionTime":"2025-10-02T09:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.847611 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.847684 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.847702 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.847730 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.847749 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:39Z","lastTransitionTime":"2025-10-02T09:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.950910 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.950993 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.951017 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.951047 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:39 crc kubenswrapper[4771]: I1002 09:38:39.951070 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:39Z","lastTransitionTime":"2025-10-02T09:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.054825 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.054892 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.054910 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.054940 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.054960 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:40Z","lastTransitionTime":"2025-10-02T09:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.159657 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.159732 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.159751 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.159780 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.159798 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:40Z","lastTransitionTime":"2025-10-02T09:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.263565 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.263645 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.263679 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.263704 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.263720 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:40Z","lastTransitionTime":"2025-10-02T09:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.366886 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.366920 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.366929 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.366943 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.366951 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:40Z","lastTransitionTime":"2025-10-02T09:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.469005 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.469071 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.469088 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.469104 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.469117 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:40Z","lastTransitionTime":"2025-10-02T09:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.572266 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.572308 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.572318 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.572333 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.572346 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:40Z","lastTransitionTime":"2025-10-02T09:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.675462 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.675551 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.675563 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.675586 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.675599 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:40Z","lastTransitionTime":"2025-10-02T09:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.680845 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.680981 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.680882 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:40 crc kubenswrapper[4771]: E1002 09:38:40.681086 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.680985 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:40 crc kubenswrapper[4771]: E1002 09:38:40.681202 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:40 crc kubenswrapper[4771]: E1002 09:38:40.681289 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:40 crc kubenswrapper[4771]: E1002 09:38:40.681388 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.778028 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.778066 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.778080 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.778095 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.778106 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:40Z","lastTransitionTime":"2025-10-02T09:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.881035 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.881082 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.881092 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.881107 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.881116 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:40Z","lastTransitionTime":"2025-10-02T09:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.983219 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.983264 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.983274 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.983290 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:40 crc kubenswrapper[4771]: I1002 09:38:40.983300 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:40Z","lastTransitionTime":"2025-10-02T09:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.125006 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.125090 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.125105 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.125155 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.125171 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:41Z","lastTransitionTime":"2025-10-02T09:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.227927 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.227984 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.227992 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.228006 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.228014 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:41Z","lastTransitionTime":"2025-10-02T09:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.331614 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.331698 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.331727 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.331764 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.331790 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:41Z","lastTransitionTime":"2025-10-02T09:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.435843 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.435913 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.435929 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.435950 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.435977 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:41Z","lastTransitionTime":"2025-10-02T09:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.538769 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.538834 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.538849 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.538877 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.538890 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:41Z","lastTransitionTime":"2025-10-02T09:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.642328 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.642390 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.642401 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.642421 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.642433 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:41Z","lastTransitionTime":"2025-10-02T09:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.745794 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.745866 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.745884 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.745912 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.745932 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:41Z","lastTransitionTime":"2025-10-02T09:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.849793 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.849872 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.849890 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.849921 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.849942 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:41Z","lastTransitionTime":"2025-10-02T09:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.953025 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.953102 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.953186 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.953225 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:41 crc kubenswrapper[4771]: I1002 09:38:41.953249 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:41Z","lastTransitionTime":"2025-10-02T09:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.056758 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.056818 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.056829 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.056847 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.056859 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:42Z","lastTransitionTime":"2025-10-02T09:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.160106 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.160197 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.160208 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.160227 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.160239 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:42Z","lastTransitionTime":"2025-10-02T09:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.263583 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.263638 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.263649 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.263667 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.263678 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:42Z","lastTransitionTime":"2025-10-02T09:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.366843 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.366923 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.366935 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.366960 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.366975 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:42Z","lastTransitionTime":"2025-10-02T09:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.470865 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.470941 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.470967 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.471004 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.471032 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:42Z","lastTransitionTime":"2025-10-02T09:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.574539 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.574610 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.574635 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.574673 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.574697 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:42Z","lastTransitionTime":"2025-10-02T09:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.678532 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.678624 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.678651 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.678686 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.678710 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:42Z","lastTransitionTime":"2025-10-02T09:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.681096 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.681210 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.681225 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.681257 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:42 crc kubenswrapper[4771]: E1002 09:38:42.681414 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:42 crc kubenswrapper[4771]: E1002 09:38:42.681529 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:42 crc kubenswrapper[4771]: E1002 09:38:42.681605 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:42 crc kubenswrapper[4771]: E1002 09:38:42.681762 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.682563 4771 scope.go:117] "RemoveContainer" containerID="777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa" Oct 02 09:38:42 crc kubenswrapper[4771]: E1002 09:38:42.682755 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4lxxg_openshift-ovn-kubernetes(2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.782740 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.782798 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.782808 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.782827 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.782841 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:42Z","lastTransitionTime":"2025-10-02T09:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.885686 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.885726 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.885737 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.885754 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.885764 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:42Z","lastTransitionTime":"2025-10-02T09:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.988425 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.988459 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.988470 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.988487 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:42 crc kubenswrapper[4771]: I1002 09:38:42.988500 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:42Z","lastTransitionTime":"2025-10-02T09:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.091341 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.091387 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.091399 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.091419 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.091433 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:43Z","lastTransitionTime":"2025-10-02T09:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.195290 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.195347 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.195366 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.195389 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.195400 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:43Z","lastTransitionTime":"2025-10-02T09:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.297576 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.297633 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.297643 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.297662 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.297673 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:43Z","lastTransitionTime":"2025-10-02T09:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.377439 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.377509 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.377519 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.377546 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.377558 4771 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:38:43Z","lastTransitionTime":"2025-10-02T09:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.440001 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2"] Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.440536 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.443724 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.445454 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.445569 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.445465 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.445569 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/822be5de-0025-4146-8469-817898f97cbd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.445677 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/822be5de-0025-4146-8469-817898f97cbd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.445707 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/822be5de-0025-4146-8469-817898f97cbd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.445738 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/822be5de-0025-4146-8469-817898f97cbd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.445813 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/822be5de-0025-4146-8469-817898f97cbd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.456325 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=25.456302237 podStartE2EDuration="25.456302237s" podCreationTimestamp="2025-10-02 09:38:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:43.455583416 +0000 UTC m=+111.103268483" watchObservedRunningTime="2025-10-02 09:38:43.456302237 +0000 UTC m=+111.103987304" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.538004 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-xh7xv" podStartSLOduration=85.537975983 podStartE2EDuration="1m25.537975983s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:43.516748968 +0000 UTC m=+111.164434035" watchObservedRunningTime="2025-10-02 09:38:43.537975983 +0000 UTC m=+111.185661050" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.538173 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-qb5kw" podStartSLOduration=85.538166948 podStartE2EDuration="1m25.538166948s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:43.53648933 +0000 UTC m=+111.184174407" watchObservedRunningTime="2025-10-02 09:38:43.538166948 +0000 UTC m=+111.185852015" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.546527 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/822be5de-0025-4146-8469-817898f97cbd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.546582 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/822be5de-0025-4146-8469-817898f97cbd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.546611 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/822be5de-0025-4146-8469-817898f97cbd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.546637 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/822be5de-0025-4146-8469-817898f97cbd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.546669 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/822be5de-0025-4146-8469-817898f97cbd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.546845 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/822be5de-0025-4146-8469-817898f97cbd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.546967 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/822be5de-0025-4146-8469-817898f97cbd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.547720 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/822be5de-0025-4146-8469-817898f97cbd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.557986 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/822be5de-0025-4146-8469-817898f97cbd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.572325 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/822be5de-0025-4146-8469-817898f97cbd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-glvx2\" (UID: \"822be5de-0025-4146-8469-817898f97cbd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.588496 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=87.588469895 podStartE2EDuration="1m27.588469895s" podCreationTimestamp="2025-10-02 09:37:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:43.56274326 +0000 UTC m=+111.210428327" watchObservedRunningTime="2025-10-02 09:38:43.588469895 +0000 UTC m=+111.236154962" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.588895 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=89.588887798 podStartE2EDuration="1m29.588887798s" podCreationTimestamp="2025-10-02 09:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:43.588314771 +0000 UTC m=+111.235999848" watchObservedRunningTime="2025-10-02 09:38:43.588887798 +0000 UTC m=+111.236572865" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.637485 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=56.637459135 podStartE2EDuration="56.637459135s" podCreationTimestamp="2025-10-02 09:37:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:43.6072675 +0000 UTC m=+111.254952587" watchObservedRunningTime="2025-10-02 09:38:43.637459135 +0000 UTC m=+111.285144202" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.669070 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-qfhxl" podStartSLOduration=86.669038359 podStartE2EDuration="1m26.669038359s" podCreationTimestamp="2025-10-02 09:37:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:43.66871549 +0000 UTC m=+111.316400567" watchObservedRunningTime="2025-10-02 09:38:43.669038359 +0000 UTC m=+111.316723436" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.670003 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wf5gr" podStartSLOduration=85.669995577 podStartE2EDuration="1m25.669995577s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:43.654737695 +0000 UTC m=+111.302422762" watchObservedRunningTime="2025-10-02 09:38:43.669995577 +0000 UTC m=+111.317680654" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.708247 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-5z94m" podStartSLOduration=86.708221834 podStartE2EDuration="1m26.708221834s" podCreationTimestamp="2025-10-02 09:37:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:43.707601337 +0000 UTC m=+111.355286414" watchObservedRunningTime="2025-10-02 09:38:43.708221834 +0000 UTC m=+111.355906911" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.721334 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podStartSLOduration=86.721308414 podStartE2EDuration="1m26.721308414s" podCreationTimestamp="2025-10-02 09:37:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:43.721143159 +0000 UTC m=+111.368828236" watchObservedRunningTime="2025-10-02 09:38:43.721308414 +0000 UTC m=+111.368993481" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.751316 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=89.751292352 podStartE2EDuration="1m29.751292352s" podCreationTimestamp="2025-10-02 09:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:43.736237876 +0000 UTC m=+111.383922953" watchObservedRunningTime="2025-10-02 09:38:43.751292352 +0000 UTC m=+111.398977419" Oct 02 09:38:43 crc kubenswrapper[4771]: I1002 09:38:43.757862 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" Oct 02 09:38:44 crc kubenswrapper[4771]: I1002 09:38:44.301234 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" event={"ID":"822be5de-0025-4146-8469-817898f97cbd","Type":"ContainerStarted","Data":"1faa22453fec7d22534d2717757d2d13369715f5ac55bd293396d38a1173381a"} Oct 02 09:38:44 crc kubenswrapper[4771]: I1002 09:38:44.301301 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" event={"ID":"822be5de-0025-4146-8469-817898f97cbd","Type":"ContainerStarted","Data":"a9e36dc31656a32f6d1cb71abc9082f765f713a6482391477e1d03e3bc340692"} Oct 02 09:38:44 crc kubenswrapper[4771]: I1002 09:38:44.318656 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glvx2" podStartSLOduration=86.318637638 podStartE2EDuration="1m26.318637638s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:44.317172735 +0000 UTC m=+111.964857832" watchObservedRunningTime="2025-10-02 09:38:44.318637638 +0000 UTC m=+111.966322705" Oct 02 09:38:44 crc kubenswrapper[4771]: I1002 09:38:44.681009 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:44 crc kubenswrapper[4771]: I1002 09:38:44.681059 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:44 crc kubenswrapper[4771]: I1002 09:38:44.681093 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:44 crc kubenswrapper[4771]: I1002 09:38:44.681038 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:44 crc kubenswrapper[4771]: E1002 09:38:44.681237 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:44 crc kubenswrapper[4771]: E1002 09:38:44.681391 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:44 crc kubenswrapper[4771]: E1002 09:38:44.681539 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:44 crc kubenswrapper[4771]: E1002 09:38:44.681740 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:46 crc kubenswrapper[4771]: I1002 09:38:46.680486 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:46 crc kubenswrapper[4771]: E1002 09:38:46.680652 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:46 crc kubenswrapper[4771]: I1002 09:38:46.680499 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:46 crc kubenswrapper[4771]: E1002 09:38:46.680755 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:46 crc kubenswrapper[4771]: I1002 09:38:46.680520 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:46 crc kubenswrapper[4771]: E1002 09:38:46.680817 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:46 crc kubenswrapper[4771]: I1002 09:38:46.680499 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:46 crc kubenswrapper[4771]: E1002 09:38:46.680874 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:48 crc kubenswrapper[4771]: I1002 09:38:48.680782 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:48 crc kubenswrapper[4771]: I1002 09:38:48.680819 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:48 crc kubenswrapper[4771]: I1002 09:38:48.680861 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:48 crc kubenswrapper[4771]: E1002 09:38:48.680879 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:48 crc kubenswrapper[4771]: I1002 09:38:48.680897 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:48 crc kubenswrapper[4771]: E1002 09:38:48.681011 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:48 crc kubenswrapper[4771]: E1002 09:38:48.681091 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:48 crc kubenswrapper[4771]: E1002 09:38:48.681530 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:50 crc kubenswrapper[4771]: I1002 09:38:50.680918 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:50 crc kubenswrapper[4771]: I1002 09:38:50.681066 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:50 crc kubenswrapper[4771]: I1002 09:38:50.681258 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:50 crc kubenswrapper[4771]: E1002 09:38:50.681774 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:50 crc kubenswrapper[4771]: E1002 09:38:50.681521 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:50 crc kubenswrapper[4771]: E1002 09:38:50.682000 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:50 crc kubenswrapper[4771]: I1002 09:38:50.681314 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:50 crc kubenswrapper[4771]: E1002 09:38:50.682222 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:52 crc kubenswrapper[4771]: I1002 09:38:52.330213 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xh7xv_bb6b65c1-83b7-4b23-abe8-c4255c6e59f8/kube-multus/1.log" Oct 02 09:38:52 crc kubenswrapper[4771]: I1002 09:38:52.330916 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xh7xv_bb6b65c1-83b7-4b23-abe8-c4255c6e59f8/kube-multus/0.log" Oct 02 09:38:52 crc kubenswrapper[4771]: I1002 09:38:52.330972 4771 generic.go:334] "Generic (PLEG): container finished" podID="bb6b65c1-83b7-4b23-abe8-c4255c6e59f8" containerID="f311d21ae303e154a68174a305b7f431b68a36d9a03fd66eaa55699ade89bcd9" exitCode=1 Oct 02 09:38:52 crc kubenswrapper[4771]: I1002 09:38:52.331009 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xh7xv" event={"ID":"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8","Type":"ContainerDied","Data":"f311d21ae303e154a68174a305b7f431b68a36d9a03fd66eaa55699ade89bcd9"} Oct 02 09:38:52 crc kubenswrapper[4771]: I1002 09:38:52.331048 4771 scope.go:117] "RemoveContainer" containerID="a45986af7547f457bdc4d2c5e9203942ccad5ad957f7d1352e5926589695a55c" Oct 02 09:38:52 crc kubenswrapper[4771]: I1002 09:38:52.331751 4771 scope.go:117] "RemoveContainer" containerID="f311d21ae303e154a68174a305b7f431b68a36d9a03fd66eaa55699ade89bcd9" Oct 02 09:38:52 crc kubenswrapper[4771]: E1002 09:38:52.332063 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-xh7xv_openshift-multus(bb6b65c1-83b7-4b23-abe8-c4255c6e59f8)\"" pod="openshift-multus/multus-xh7xv" podUID="bb6b65c1-83b7-4b23-abe8-c4255c6e59f8" Oct 02 09:38:52 crc kubenswrapper[4771]: I1002 09:38:52.680274 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:52 crc kubenswrapper[4771]: I1002 09:38:52.680357 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:52 crc kubenswrapper[4771]: I1002 09:38:52.680529 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:52 crc kubenswrapper[4771]: E1002 09:38:52.680524 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:52 crc kubenswrapper[4771]: I1002 09:38:52.680638 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:52 crc kubenswrapper[4771]: E1002 09:38:52.680901 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:52 crc kubenswrapper[4771]: E1002 09:38:52.681182 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:52 crc kubenswrapper[4771]: E1002 09:38:52.681345 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:53 crc kubenswrapper[4771]: I1002 09:38:53.334756 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xh7xv_bb6b65c1-83b7-4b23-abe8-c4255c6e59f8/kube-multus/1.log" Oct 02 09:38:53 crc kubenswrapper[4771]: E1002 09:38:53.535863 4771 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 02 09:38:53 crc kubenswrapper[4771]: I1002 09:38:53.681980 4771 scope.go:117] "RemoveContainer" containerID="777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa" Oct 02 09:38:53 crc kubenswrapper[4771]: E1002 09:38:53.908395 4771 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:38:54 crc kubenswrapper[4771]: I1002 09:38:54.340075 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/3.log" Oct 02 09:38:54 crc kubenswrapper[4771]: I1002 09:38:54.342490 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerStarted","Data":"43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251"} Oct 02 09:38:54 crc kubenswrapper[4771]: I1002 09:38:54.343025 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:38:54 crc kubenswrapper[4771]: I1002 09:38:54.371513 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podStartSLOduration=96.371494707 podStartE2EDuration="1m36.371494707s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:54.369563861 +0000 UTC m=+122.017248928" watchObservedRunningTime="2025-10-02 09:38:54.371494707 +0000 UTC m=+122.019179774" Oct 02 09:38:54 crc kubenswrapper[4771]: I1002 09:38:54.680761 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:54 crc kubenswrapper[4771]: I1002 09:38:54.680828 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:54 crc kubenswrapper[4771]: I1002 09:38:54.680834 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:54 crc kubenswrapper[4771]: I1002 09:38:54.680787 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:54 crc kubenswrapper[4771]: E1002 09:38:54.680954 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:54 crc kubenswrapper[4771]: E1002 09:38:54.681059 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:54 crc kubenswrapper[4771]: E1002 09:38:54.681321 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:54 crc kubenswrapper[4771]: E1002 09:38:54.681269 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:54 crc kubenswrapper[4771]: I1002 09:38:54.685574 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zp7kh"] Oct 02 09:38:55 crc kubenswrapper[4771]: I1002 09:38:55.345600 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:55 crc kubenswrapper[4771]: E1002 09:38:55.346013 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:56 crc kubenswrapper[4771]: I1002 09:38:56.680897 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:56 crc kubenswrapper[4771]: I1002 09:38:56.680913 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:56 crc kubenswrapper[4771]: E1002 09:38:56.681730 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:56 crc kubenswrapper[4771]: E1002 09:38:56.681788 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:56 crc kubenswrapper[4771]: I1002 09:38:56.680983 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:56 crc kubenswrapper[4771]: E1002 09:38:56.681881 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:56 crc kubenswrapper[4771]: I1002 09:38:56.681015 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:56 crc kubenswrapper[4771]: E1002 09:38:56.682232 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:58 crc kubenswrapper[4771]: I1002 09:38:58.680880 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:38:58 crc kubenswrapper[4771]: I1002 09:38:58.680886 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:38:58 crc kubenswrapper[4771]: E1002 09:38:58.681019 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:38:58 crc kubenswrapper[4771]: I1002 09:38:58.680907 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:38:58 crc kubenswrapper[4771]: E1002 09:38:58.681103 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:38:58 crc kubenswrapper[4771]: I1002 09:38:58.680891 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:38:58 crc kubenswrapper[4771]: E1002 09:38:58.681207 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:38:58 crc kubenswrapper[4771]: E1002 09:38:58.681261 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:38:58 crc kubenswrapper[4771]: E1002 09:38:58.910695 4771 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:39:00 crc kubenswrapper[4771]: I1002 09:39:00.680327 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:39:00 crc kubenswrapper[4771]: I1002 09:39:00.680480 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:39:00 crc kubenswrapper[4771]: E1002 09:39:00.680554 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:39:00 crc kubenswrapper[4771]: I1002 09:39:00.680349 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:39:00 crc kubenswrapper[4771]: I1002 09:39:00.680374 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:39:00 crc kubenswrapper[4771]: E1002 09:39:00.680722 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:39:00 crc kubenswrapper[4771]: E1002 09:39:00.681049 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:39:00 crc kubenswrapper[4771]: E1002 09:39:00.681208 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:39:02 crc kubenswrapper[4771]: I1002 09:39:02.680190 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:39:02 crc kubenswrapper[4771]: I1002 09:39:02.680190 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:39:02 crc kubenswrapper[4771]: E1002 09:39:02.680343 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:39:02 crc kubenswrapper[4771]: I1002 09:39:02.680209 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:39:02 crc kubenswrapper[4771]: I1002 09:39:02.680190 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:39:02 crc kubenswrapper[4771]: E1002 09:39:02.680433 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:39:02 crc kubenswrapper[4771]: E1002 09:39:02.680486 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:39:02 crc kubenswrapper[4771]: E1002 09:39:02.680397 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:39:03 crc kubenswrapper[4771]: E1002 09:39:03.911346 4771 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:39:04 crc kubenswrapper[4771]: I1002 09:39:04.680592 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:39:04 crc kubenswrapper[4771]: E1002 09:39:04.680828 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:39:04 crc kubenswrapper[4771]: I1002 09:39:04.680848 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:39:04 crc kubenswrapper[4771]: I1002 09:39:04.680866 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:39:04 crc kubenswrapper[4771]: I1002 09:39:04.681227 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:39:04 crc kubenswrapper[4771]: I1002 09:39:04.681249 4771 scope.go:117] "RemoveContainer" containerID="f311d21ae303e154a68174a305b7f431b68a36d9a03fd66eaa55699ade89bcd9" Oct 02 09:39:04 crc kubenswrapper[4771]: E1002 09:39:04.681240 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:39:04 crc kubenswrapper[4771]: E1002 09:39:04.681472 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:39:04 crc kubenswrapper[4771]: E1002 09:39:04.681566 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:39:05 crc kubenswrapper[4771]: I1002 09:39:05.377316 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xh7xv_bb6b65c1-83b7-4b23-abe8-c4255c6e59f8/kube-multus/1.log" Oct 02 09:39:05 crc kubenswrapper[4771]: I1002 09:39:05.377648 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xh7xv" event={"ID":"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8","Type":"ContainerStarted","Data":"c35db692e359d56ccbb79957ddfb37fbf56b3ad4e46093fa55f540da775d2441"} Oct 02 09:39:06 crc kubenswrapper[4771]: I1002 09:39:06.680591 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:39:06 crc kubenswrapper[4771]: I1002 09:39:06.680663 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:39:06 crc kubenswrapper[4771]: I1002 09:39:06.680757 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:39:06 crc kubenswrapper[4771]: I1002 09:39:06.680949 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:39:06 crc kubenswrapper[4771]: E1002 09:39:06.680934 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:39:06 crc kubenswrapper[4771]: E1002 09:39:06.681050 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:39:06 crc kubenswrapper[4771]: E1002 09:39:06.681118 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:39:06 crc kubenswrapper[4771]: E1002 09:39:06.681225 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:39:08 crc kubenswrapper[4771]: I1002 09:39:08.680700 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:39:08 crc kubenswrapper[4771]: I1002 09:39:08.680829 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:39:08 crc kubenswrapper[4771]: I1002 09:39:08.680897 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:39:08 crc kubenswrapper[4771]: E1002 09:39:08.680930 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:39:08 crc kubenswrapper[4771]: I1002 09:39:08.681061 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:39:08 crc kubenswrapper[4771]: E1002 09:39:08.681240 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:39:08 crc kubenswrapper[4771]: E1002 09:39:08.681357 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:39:08 crc kubenswrapper[4771]: E1002 09:39:08.681437 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zp7kh" podUID="20d5431d-4595-4b96-a8b8-8953e3dffb53" Oct 02 09:39:10 crc kubenswrapper[4771]: I1002 09:39:10.680987 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:39:10 crc kubenswrapper[4771]: I1002 09:39:10.680999 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:39:10 crc kubenswrapper[4771]: I1002 09:39:10.681298 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:39:10 crc kubenswrapper[4771]: I1002 09:39:10.680997 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:39:10 crc kubenswrapper[4771]: I1002 09:39:10.686279 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 02 09:39:10 crc kubenswrapper[4771]: I1002 09:39:10.688247 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 02 09:39:10 crc kubenswrapper[4771]: I1002 09:39:10.688592 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 02 09:39:10 crc kubenswrapper[4771]: I1002 09:39:10.689244 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 02 09:39:10 crc kubenswrapper[4771]: I1002 09:39:10.690743 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 02 09:39:10 crc kubenswrapper[4771]: I1002 09:39:10.691572 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.784547 4771 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.827310 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sv9ls"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.827887 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.828993 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-68c2v"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.830573 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.830674 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.831824 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.832088 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.832611 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.832802 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.832851 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.833009 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.833070 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.834337 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.834599 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.845231 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.845721 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.845875 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.846163 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.846246 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.846278 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.846722 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.847035 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.847340 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.847444 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.847550 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.847605 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.847960 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.848121 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.848569 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.850263 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.853429 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8dj62"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.853892 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.856201 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-vk4pw"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.856857 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vk4pw" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.856993 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.857417 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.859055 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.859257 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.859215 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.865801 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.866315 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.869214 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-p65v7"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.870059 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.870864 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.871103 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.874737 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.875328 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.875566 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.875725 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.876092 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.876224 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.876994 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.877103 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.877425 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.879187 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-r664s"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.879392 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.880005 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.880367 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.884686 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-gdzf9"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.885442 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.886434 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.886551 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.886646 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.886807 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.886832 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.886900 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.886938 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.886952 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.887195 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.887959 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.888079 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.888292 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.888431 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.892065 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.893326 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.894397 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.894740 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.894976 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.895089 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.898293 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.898590 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.898759 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.898994 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.899175 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.899293 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.899501 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.899174 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.899752 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.900119 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nbmlz"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.900849 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.901077 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8dj62"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.903383 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.904091 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.904501 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.904707 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.904902 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.906219 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sv9ls"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.907420 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.907553 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.907680 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.908149 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.908332 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.908656 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.910394 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.910631 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.918379 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.919012 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.925592 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.926774 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vk4pw"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.926823 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.927843 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.929290 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.929383 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.930574 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.930751 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.931205 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.931551 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.931883 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.931983 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.932270 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.932481 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.940366 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.940572 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.941908 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.942406 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.945317 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.945560 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.968077 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jb4nn"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.969324 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.969727 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vz4s6"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.971016 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.971185 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.971247 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jb4nn" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.972830 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.973506 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.972903 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.974015 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p2srt"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.982487 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-gdzf9"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.982516 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.982531 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-nfr9d"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.982985 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.983313 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.975504 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.975551 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.978997 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.979271 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.979551 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.983723 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.983778 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.983970 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nfr9d" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.984114 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.988206 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.989243 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.990149 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.991343 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.991823 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl"] Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.997363 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.998986 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.999398 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 02 09:39:13 crc kubenswrapper[4771]: I1002 09:39:13.999784 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.000698 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.001804 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.004034 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.004685 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.006336 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-p65v7"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.006785 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-service-ca-bundle\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.006844 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-audit\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.006868 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.006890 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-config\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.006913 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-serving-cert\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.006934 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-client-ca\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.006954 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76c97829-08c2-4e93-ae1f-4788dca51fc4-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pxjp8\" (UID: \"76c97829-08c2-4e93-ae1f-4788dca51fc4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.006978 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/961e712e-5760-4436-9be7-6bb6be2f8cf3-audit-dir\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.006998 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-serving-cert\") pod \"route-controller-manager-6576b87f9c-g92wn\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007044 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lph2s\" (UniqueName: \"kubernetes.io/projected/e0904a9e-81f3-4849-9f7d-90e2485a218a-kube-api-access-lph2s\") pod \"cluster-samples-operator-665b6dd947-rlhp2\" (UID: \"e0904a9e-81f3-4849-9f7d-90e2485a218a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007097 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-config\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007202 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007199 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c65812c5-b7ed-434b-bdef-bbbe26fe5c69-auth-proxy-config\") pod \"machine-approver-56656f9798-6zk5s\" (UID: \"c65812c5-b7ed-434b-bdef-bbbe26fe5c69\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007282 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-config\") pod \"route-controller-manager-6576b87f9c-g92wn\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007328 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-config\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007355 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/671d32ff-3e87-49bd-a4ec-60a495671813-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zlrjz\" (UID: \"671d32ff-3e87-49bd-a4ec-60a495671813\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007377 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4qfx\" (UniqueName: \"kubernetes.io/projected/76c97829-08c2-4e93-ae1f-4788dca51fc4-kube-api-access-b4qfx\") pod \"openshift-apiserver-operator-796bbdcf4f-pxjp8\" (UID: \"76c97829-08c2-4e93-ae1f-4788dca51fc4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007408 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fefa105a-53b5-45a0-ac11-f3078d78ae59-audit-policies\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007426 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007496 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/961e712e-5760-4436-9be7-6bb6be2f8cf3-etcd-client\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007590 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007621 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tntj5\" (UniqueName: \"kubernetes.io/projected/e3d49353-2942-437a-bfbb-2da3e5df246d-kube-api-access-tntj5\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007646 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007666 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e62d3902-9b6c-48ec-af1f-4ee9823ccc12-webhook-cert\") pod \"packageserver-d55dfcdfc-xtdpd\" (UID: \"e62d3902-9b6c-48ec-af1f-4ee9823ccc12\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007704 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1cfc1793-db10-47a8-a152-0c86cc82be4c-images\") pod \"machine-api-operator-5694c8668f-68c2v\" (UID: \"1cfc1793-db10-47a8-a152-0c86cc82be4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007726 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76c97829-08c2-4e93-ae1f-4788dca51fc4-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pxjp8\" (UID: \"76c97829-08c2-4e93-ae1f-4788dca51fc4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007742 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c65812c5-b7ed-434b-bdef-bbbe26fe5c69-config\") pod \"machine-approver-56656f9798-6zk5s\" (UID: \"c65812c5-b7ed-434b-bdef-bbbe26fe5c69\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007810 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fefa105a-53b5-45a0-ac11-f3078d78ae59-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007838 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007901 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjqfx\" (UniqueName: \"kubernetes.io/projected/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-kube-api-access-wjqfx\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007919 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/961e712e-5760-4436-9be7-6bb6be2f8cf3-encryption-config\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007969 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1366a10-f981-4c9a-aecd-415273b5b714-config\") pod \"console-operator-58897d9998-8dj62\" (UID: \"d1366a10-f981-4c9a-aecd-415273b5b714\") " pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.007987 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/961e712e-5760-4436-9be7-6bb6be2f8cf3-node-pullsecrets\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008029 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/961e712e-5760-4436-9be7-6bb6be2f8cf3-serving-cert\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008061 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4gln\" (UniqueName: \"kubernetes.io/projected/961e712e-5760-4436-9be7-6bb6be2f8cf3-kube-api-access-r4gln\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008226 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008252 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvlgw\" (UniqueName: \"kubernetes.io/projected/c65812c5-b7ed-434b-bdef-bbbe26fe5c69-kube-api-access-jvlgw\") pod \"machine-approver-56656f9798-6zk5s\" (UID: \"c65812c5-b7ed-434b-bdef-bbbe26fe5c69\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008271 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mcbd\" (UniqueName: \"kubernetes.io/projected/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-kube-api-access-8mcbd\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008387 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e0904a9e-81f3-4849-9f7d-90e2485a218a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rlhp2\" (UID: \"e0904a9e-81f3-4849-9f7d-90e2485a218a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008408 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/671d32ff-3e87-49bd-a4ec-60a495671813-serving-cert\") pod \"openshift-config-operator-7777fb866f-zlrjz\" (UID: \"671d32ff-3e87-49bd-a4ec-60a495671813\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008428 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008548 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fefa105a-53b5-45a0-ac11-f3078d78ae59-encryption-config\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008583 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-oauth-config\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008716 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008749 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-client-ca\") pod \"route-controller-manager-6576b87f9c-g92wn\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008878 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmhkn\" (UniqueName: \"kubernetes.io/projected/671d32ff-3e87-49bd-a4ec-60a495671813-kube-api-access-vmhkn\") pod \"openshift-config-operator-7777fb866f-zlrjz\" (UID: \"671d32ff-3e87-49bd-a4ec-60a495671813\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.008909 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fefa105a-53b5-45a0-ac11-f3078d78ae59-etcd-client\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.009058 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-config\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.009086 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e3d49353-2942-437a-bfbb-2da3e5df246d-audit-dir\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.009218 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.009247 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fefa105a-53b5-45a0-ac11-f3078d78ae59-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.009373 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-service-ca\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.009525 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.009600 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c65812c5-b7ed-434b-bdef-bbbe26fe5c69-machine-approver-tls\") pod \"machine-approver-56656f9798-6zk5s\" (UID: \"c65812c5-b7ed-434b-bdef-bbbe26fe5c69\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.009639 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e62d3902-9b6c-48ec-af1f-4ee9823ccc12-tmpfs\") pod \"packageserver-d55dfcdfc-xtdpd\" (UID: \"e62d3902-9b6c-48ec-af1f-4ee9823ccc12\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.009719 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1cfc1793-db10-47a8-a152-0c86cc82be4c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-68c2v\" (UID: \"1cfc1793-db10-47a8-a152-0c86cc82be4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.009768 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f99ns\" (UniqueName: \"kubernetes.io/projected/1cfc1793-db10-47a8-a152-0c86cc82be4c-kube-api-access-f99ns\") pod \"machine-api-operator-5694c8668f-68c2v\" (UID: \"1cfc1793-db10-47a8-a152-0c86cc82be4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.009862 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d1366a10-f981-4c9a-aecd-415273b5b714-trusted-ca\") pod \"console-operator-58897d9998-8dj62\" (UID: \"d1366a10-f981-4c9a-aecd-415273b5b714\") " pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.009891 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-549tf\" (UniqueName: \"kubernetes.io/projected/e62d3902-9b6c-48ec-af1f-4ee9823ccc12-kube-api-access-549tf\") pod \"packageserver-d55dfcdfc-xtdpd\" (UID: \"e62d3902-9b6c-48ec-af1f-4ee9823ccc12\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.009957 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e62d3902-9b6c-48ec-af1f-4ee9823ccc12-apiservice-cert\") pod \"packageserver-d55dfcdfc-xtdpd\" (UID: \"e62d3902-9b6c-48ec-af1f-4ee9823ccc12\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010002 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cfc1793-db10-47a8-a152-0c86cc82be4c-config\") pod \"machine-api-operator-5694c8668f-68c2v\" (UID: \"1cfc1793-db10-47a8-a152-0c86cc82be4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010034 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgmmt\" (UniqueName: \"kubernetes.io/projected/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-kube-api-access-bgmmt\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010052 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fefa105a-53b5-45a0-ac11-f3078d78ae59-serving-cert\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010083 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1366a10-f981-4c9a-aecd-415273b5b714-serving-cert\") pod \"console-operator-58897d9998-8dj62\" (UID: \"d1366a10-f981-4c9a-aecd-415273b5b714\") " pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010103 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-oauth-serving-cert\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010117 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-etcd-serving-ca\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010295 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz9tq\" (UniqueName: \"kubernetes.io/projected/77e60bc8-1d3d-4465-8dea-d7840a681b32-kube-api-access-gz9tq\") pod \"downloads-7954f5f757-vk4pw\" (UID: \"77e60bc8-1d3d-4465-8dea-d7840a681b32\") " pod="openshift-console/downloads-7954f5f757-vk4pw" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010361 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtcgt\" (UniqueName: \"kubernetes.io/projected/d1366a10-f981-4c9a-aecd-415273b5b714-kube-api-access-mtcgt\") pod \"console-operator-58897d9998-8dj62\" (UID: \"d1366a10-f981-4c9a-aecd-415273b5b714\") " pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010409 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-audit-policies\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010437 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzxh7\" (UniqueName: \"kubernetes.io/projected/fefa105a-53b5-45a0-ac11-f3078d78ae59-kube-api-access-vzxh7\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010461 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-image-import-ca\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010482 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49xvl\" (UniqueName: \"kubernetes.io/projected/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-kube-api-access-49xvl\") pod \"route-controller-manager-6576b87f9c-g92wn\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010507 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-serving-cert\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010534 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-trusted-ca-bundle\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010565 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010604 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010626 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010650 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010670 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fefa105a-53b5-45a0-ac11-f3078d78ae59-audit-dir\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.010693 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-serving-cert\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.015564 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.022766 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.023427 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.047729 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.047991 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-r664s"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.049554 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.052033 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cx2j5"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.053634 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.054766 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.056155 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.058761 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.059447 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.060237 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.060984 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.061713 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.063789 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7prrf"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.064355 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-l4x98"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.064738 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.064782 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7prrf" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.064754 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.065517 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.067527 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-z7f8f"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.068276 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.070830 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-n6z8c"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.071147 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.071346 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.072087 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.072380 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-n6z8c" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.074720 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.076812 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.078269 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.079110 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.079535 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vz4s6"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.080780 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.082090 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.084052 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-68c2v"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.085426 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.086798 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.088445 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-n6z8c"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.089630 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.090444 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.092876 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.094897 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.096675 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jb4nn"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.098377 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-nfr9d"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.100300 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nbmlz"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.101894 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p2srt"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.103779 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-l4x98"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.105043 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.107656 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cx2j5"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.110277 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.110356 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111264 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/961e712e-5760-4436-9be7-6bb6be2f8cf3-etcd-client\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111292 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/671d32ff-3e87-49bd-a4ec-60a495671813-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zlrjz\" (UID: \"671d32ff-3e87-49bd-a4ec-60a495671813\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111315 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4qfx\" (UniqueName: \"kubernetes.io/projected/76c97829-08c2-4e93-ae1f-4788dca51fc4-kube-api-access-b4qfx\") pod \"openshift-apiserver-operator-796bbdcf4f-pxjp8\" (UID: \"76c97829-08c2-4e93-ae1f-4788dca51fc4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111333 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fefa105a-53b5-45a0-ac11-f3078d78ae59-audit-policies\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111349 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111373 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-serving-cert\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111390 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzvp8\" (UniqueName: \"kubernetes.io/projected/a833cc23-a6bc-4d07-95bf-d745791879d6-kube-api-access-kzvp8\") pod \"olm-operator-6b444d44fb-4wzs8\" (UID: \"a833cc23-a6bc-4d07-95bf-d745791879d6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111414 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111432 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d039518e-be81-4d51-9798-7043a3b653e0-metrics-tls\") pod \"ingress-operator-5b745b69d9-58fw2\" (UID: \"d039518e-be81-4d51-9798-7043a3b653e0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111451 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tntj5\" (UniqueName: \"kubernetes.io/projected/e3d49353-2942-437a-bfbb-2da3e5df246d-kube-api-access-tntj5\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111509 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz4lx\" (UniqueName: \"kubernetes.io/projected/d039518e-be81-4d51-9798-7043a3b653e0-kube-api-access-wz4lx\") pod \"ingress-operator-5b745b69d9-58fw2\" (UID: \"d039518e-be81-4d51-9798-7043a3b653e0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111541 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88b88dc4-a78f-489f-b970-f70828c68f6d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6n2dp\" (UID: \"88b88dc4-a78f-489f-b970-f70828c68f6d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111562 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gzfw\" (UniqueName: \"kubernetes.io/projected/9f20b7d3-21f9-4a06-a160-df1bd47f389d-kube-api-access-6gzfw\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vcxd\" (UID: \"9f20b7d3-21f9-4a06-a160-df1bd47f389d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111618 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111637 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e62d3902-9b6c-48ec-af1f-4ee9823ccc12-webhook-cert\") pod \"packageserver-d55dfcdfc-xtdpd\" (UID: \"e62d3902-9b6c-48ec-af1f-4ee9823ccc12\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111655 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1cfc1793-db10-47a8-a152-0c86cc82be4c-images\") pod \"machine-api-operator-5694c8668f-68c2v\" (UID: \"1cfc1793-db10-47a8-a152-0c86cc82be4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111670 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76c97829-08c2-4e93-ae1f-4788dca51fc4-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pxjp8\" (UID: \"76c97829-08c2-4e93-ae1f-4788dca51fc4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111686 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c65812c5-b7ed-434b-bdef-bbbe26fe5c69-config\") pod \"machine-approver-56656f9798-6zk5s\" (UID: \"c65812c5-b7ed-434b-bdef-bbbe26fe5c69\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111706 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-etcd-service-ca\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111720 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f20b7d3-21f9-4a06-a160-df1bd47f389d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vcxd\" (UID: \"9f20b7d3-21f9-4a06-a160-df1bd47f389d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111755 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fefa105a-53b5-45a0-ac11-f3078d78ae59-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111798 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/671d32ff-3e87-49bd-a4ec-60a495671813-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zlrjz\" (UID: \"671d32ff-3e87-49bd-a4ec-60a495671813\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111843 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111897 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qqm8\" (UniqueName: \"kubernetes.io/projected/c8cb8248-6e33-49aa-9422-f2e14c0e9015-kube-api-access-9qqm8\") pod \"cluster-image-registry-operator-dc59b4c8b-h8j9j\" (UID: \"c8cb8248-6e33-49aa-9422-f2e14c0e9015\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111971 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjqfx\" (UniqueName: \"kubernetes.io/projected/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-kube-api-access-wjqfx\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111983 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fefa105a-53b5-45a0-ac11-f3078d78ae59-audit-policies\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.111993 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/961e712e-5760-4436-9be7-6bb6be2f8cf3-encryption-config\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112083 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4gln\" (UniqueName: \"kubernetes.io/projected/961e712e-5760-4436-9be7-6bb6be2f8cf3-kube-api-access-r4gln\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112144 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1366a10-f981-4c9a-aecd-415273b5b714-config\") pod \"console-operator-58897d9998-8dj62\" (UID: \"d1366a10-f981-4c9a-aecd-415273b5b714\") " pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112279 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/961e712e-5760-4436-9be7-6bb6be2f8cf3-node-pullsecrets\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112314 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/961e712e-5760-4436-9be7-6bb6be2f8cf3-serving-cert\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112339 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112365 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvlgw\" (UniqueName: \"kubernetes.io/projected/c65812c5-b7ed-434b-bdef-bbbe26fe5c69-kube-api-access-jvlgw\") pod \"machine-approver-56656f9798-6zk5s\" (UID: \"c65812c5-b7ed-434b-bdef-bbbe26fe5c69\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112415 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8cb8248-6e33-49aa-9422-f2e14c0e9015-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h8j9j\" (UID: \"c8cb8248-6e33-49aa-9422-f2e14c0e9015\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112438 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/671d32ff-3e87-49bd-a4ec-60a495671813-serving-cert\") pod \"openshift-config-operator-7777fb866f-zlrjz\" (UID: \"671d32ff-3e87-49bd-a4ec-60a495671813\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112461 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112486 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mcbd\" (UniqueName: \"kubernetes.io/projected/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-kube-api-access-8mcbd\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112528 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e0904a9e-81f3-4849-9f7d-90e2485a218a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rlhp2\" (UID: \"e0904a9e-81f3-4849-9f7d-90e2485a218a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112602 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0ffcb835-7bd1-4cb3-b7cf-354595aa0b27-proxy-tls\") pod \"machine-config-controller-84d6567774-r8c2x\" (UID: \"0ffcb835-7bd1-4cb3-b7cf-354595aa0b27\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112630 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9fd48e0d-81e5-49ed-8ced-4c382f96f831-metrics-tls\") pod \"dns-operator-744455d44c-jb4nn\" (UID: \"9fd48e0d-81e5-49ed-8ced-4c382f96f831\") " pod="openshift-dns-operator/dns-operator-744455d44c-jb4nn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112669 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmhkn\" (UniqueName: \"kubernetes.io/projected/671d32ff-3e87-49bd-a4ec-60a495671813-kube-api-access-vmhkn\") pod \"openshift-config-operator-7777fb866f-zlrjz\" (UID: \"671d32ff-3e87-49bd-a4ec-60a495671813\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112692 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fefa105a-53b5-45a0-ac11-f3078d78ae59-etcd-client\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.112841 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fefa105a-53b5-45a0-ac11-f3078d78ae59-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.113208 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fefa105a-53b5-45a0-ac11-f3078d78ae59-encryption-config\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.113717 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.116081 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1366a10-f981-4c9a-aecd-415273b5b714-config\") pod \"console-operator-58897d9998-8dj62\" (UID: \"d1366a10-f981-4c9a-aecd-415273b5b714\") " pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.116144 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/961e712e-5760-4436-9be7-6bb6be2f8cf3-node-pullsecrets\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.116216 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-oauth-config\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.116294 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.116480 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c65812c5-b7ed-434b-bdef-bbbe26fe5c69-config\") pod \"machine-approver-56656f9798-6zk5s\" (UID: \"c65812c5-b7ed-434b-bdef-bbbe26fe5c69\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.116626 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1cfc1793-db10-47a8-a152-0c86cc82be4c-images\") pod \"machine-api-operator-5694c8668f-68c2v\" (UID: \"1cfc1793-db10-47a8-a152-0c86cc82be4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.116321 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-client-ca\") pod \"route-controller-manager-6576b87f9c-g92wn\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.117000 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.117482 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fefa105a-53b5-45a0-ac11-f3078d78ae59-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.117540 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-config\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.117563 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e3d49353-2942-437a-bfbb-2da3e5df246d-audit-dir\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.117605 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.117626 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f20b7d3-21f9-4a06-a160-df1bd47f389d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vcxd\" (UID: \"9f20b7d3-21f9-4a06-a160-df1bd47f389d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.117687 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-client-ca\") pod \"route-controller-manager-6576b87f9c-g92wn\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.117718 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3217c27-f7c4-4b0f-9503-1968ce24eaba-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rrb82\" (UID: \"b3217c27-f7c4-4b0f-9503-1968ce24eaba\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.117753 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e3d49353-2942-437a-bfbb-2da3e5df246d-audit-dir\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.117824 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-service-ca\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.117853 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.117910 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz2v2\" (UniqueName: \"kubernetes.io/projected/19247691-b8da-4344-91e1-57bc628fca0a-kube-api-access-rz2v2\") pod \"migrator-59844c95c7-nfr9d\" (UID: \"19247691-b8da-4344-91e1-57bc628fca0a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nfr9d" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.117931 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkkzw\" (UniqueName: \"kubernetes.io/projected/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-kube-api-access-pkkzw\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.118289 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fefa105a-53b5-45a0-ac11-f3078d78ae59-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.118334 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e62d3902-9b6c-48ec-af1f-4ee9823ccc12-tmpfs\") pod \"packageserver-d55dfcdfc-xtdpd\" (UID: \"e62d3902-9b6c-48ec-af1f-4ee9823ccc12\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.118370 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c65812c5-b7ed-434b-bdef-bbbe26fe5c69-machine-approver-tls\") pod \"machine-approver-56656f9798-6zk5s\" (UID: \"c65812c5-b7ed-434b-bdef-bbbe26fe5c69\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.118523 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.118609 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-config\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.118677 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e62d3902-9b6c-48ec-af1f-4ee9823ccc12-tmpfs\") pod \"packageserver-d55dfcdfc-xtdpd\" (UID: \"e62d3902-9b6c-48ec-af1f-4ee9823ccc12\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.118677 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f99ns\" (UniqueName: \"kubernetes.io/projected/1cfc1793-db10-47a8-a152-0c86cc82be4c-kube-api-access-f99ns\") pod \"machine-api-operator-5694c8668f-68c2v\" (UID: \"1cfc1793-db10-47a8-a152-0c86cc82be4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.118838 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1cfc1793-db10-47a8-a152-0c86cc82be4c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-68c2v\" (UID: \"1cfc1793-db10-47a8-a152-0c86cc82be4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.118878 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a833cc23-a6bc-4d07-95bf-d745791879d6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4wzs8\" (UID: \"a833cc23-a6bc-4d07-95bf-d745791879d6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.118950 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c8cb8248-6e33-49aa-9422-f2e14c0e9015-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h8j9j\" (UID: \"c8cb8248-6e33-49aa-9422-f2e14c0e9015\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119030 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d1366a10-f981-4c9a-aecd-415273b5b714-trusted-ca\") pod \"console-operator-58897d9998-8dj62\" (UID: \"d1366a10-f981-4c9a-aecd-415273b5b714\") " pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119074 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-549tf\" (UniqueName: \"kubernetes.io/projected/e62d3902-9b6c-48ec-af1f-4ee9823ccc12-kube-api-access-549tf\") pod \"packageserver-d55dfcdfc-xtdpd\" (UID: \"e62d3902-9b6c-48ec-af1f-4ee9823ccc12\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119187 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a833cc23-a6bc-4d07-95bf-d745791879d6-srv-cert\") pod \"olm-operator-6b444d44fb-4wzs8\" (UID: \"a833cc23-a6bc-4d07-95bf-d745791879d6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119326 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgmmt\" (UniqueName: \"kubernetes.io/projected/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-kube-api-access-bgmmt\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119355 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fefa105a-53b5-45a0-ac11-f3078d78ae59-serving-cert\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119375 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e62d3902-9b6c-48ec-af1f-4ee9823ccc12-apiservice-cert\") pod \"packageserver-d55dfcdfc-xtdpd\" (UID: \"e62d3902-9b6c-48ec-af1f-4ee9823ccc12\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119398 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cfc1793-db10-47a8-a152-0c86cc82be4c-config\") pod \"machine-api-operator-5694c8668f-68c2v\" (UID: \"1cfc1793-db10-47a8-a152-0c86cc82be4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119427 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb4nc\" (UniqueName: \"kubernetes.io/projected/0ffcb835-7bd1-4cb3-b7cf-354595aa0b27-kube-api-access-jb4nc\") pod \"machine-config-controller-84d6567774-r8c2x\" (UID: \"0ffcb835-7bd1-4cb3-b7cf-354595aa0b27\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119451 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-config\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119507 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1366a10-f981-4c9a-aecd-415273b5b714-serving-cert\") pod \"console-operator-58897d9998-8dj62\" (UID: \"d1366a10-f981-4c9a-aecd-415273b5b714\") " pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119532 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-oauth-serving-cert\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119555 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-etcd-serving-ca\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119591 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz9tq\" (UniqueName: \"kubernetes.io/projected/77e60bc8-1d3d-4465-8dea-d7840a681b32-kube-api-access-gz9tq\") pod \"downloads-7954f5f757-vk4pw\" (UID: \"77e60bc8-1d3d-4465-8dea-d7840a681b32\") " pod="openshift-console/downloads-7954f5f757-vk4pw" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119628 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtcgt\" (UniqueName: \"kubernetes.io/projected/d1366a10-f981-4c9a-aecd-415273b5b714-kube-api-access-mtcgt\") pod \"console-operator-58897d9998-8dj62\" (UID: \"d1366a10-f981-4c9a-aecd-415273b5b714\") " pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119661 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-audit-policies\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119685 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzxh7\" (UniqueName: \"kubernetes.io/projected/fefa105a-53b5-45a0-ac11-f3078d78ae59-kube-api-access-vzxh7\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119711 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-serving-cert\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119735 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-trusted-ca-bundle\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.121500 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/961e712e-5760-4436-9be7-6bb6be2f8cf3-encryption-config\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.122082 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d1366a10-f981-4c9a-aecd-415273b5b714-trusted-ca\") pod \"console-operator-58897d9998-8dj62\" (UID: \"d1366a10-f981-4c9a-aecd-415273b5b714\") " pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.122572 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.122938 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.123210 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fefa105a-53b5-45a0-ac11-f3078d78ae59-etcd-client\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.123796 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-audit-policies\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.124309 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cfc1793-db10-47a8-a152-0c86cc82be4c-config\") pod \"machine-api-operator-5694c8668f-68c2v\" (UID: \"1cfc1793-db10-47a8-a152-0c86cc82be4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.124752 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76c97829-08c2-4e93-ae1f-4788dca51fc4-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pxjp8\" (UID: \"76c97829-08c2-4e93-ae1f-4788dca51fc4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.125141 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-trusted-ca-bundle\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.125315 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e0904a9e-81f3-4849-9f7d-90e2485a218a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rlhp2\" (UID: \"e0904a9e-81f3-4849-9f7d-90e2485a218a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.125564 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.125619 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pr9z4"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.125934 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-oauth-serving-cert\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.125937 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fefa105a-53b5-45a0-ac11-f3078d78ae59-serving-cert\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.126226 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.126444 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e62d3902-9b6c-48ec-af1f-4ee9823ccc12-webhook-cert\") pod \"packageserver-d55dfcdfc-xtdpd\" (UID: \"e62d3902-9b6c-48ec-af1f-4ee9823ccc12\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.127418 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-oauth-config\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.127591 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/671d32ff-3e87-49bd-a4ec-60a495671813-serving-cert\") pod \"openshift-config-operator-7777fb866f-zlrjz\" (UID: \"671d32ff-3e87-49bd-a4ec-60a495671813\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.127699 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-zpnmm"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.127903 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c65812c5-b7ed-434b-bdef-bbbe26fe5c69-machine-approver-tls\") pod \"machine-approver-56656f9798-6zk5s\" (UID: \"c65812c5-b7ed-434b-bdef-bbbe26fe5c69\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.125579 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fefa105a-53b5-45a0-ac11-f3078d78ae59-encryption-config\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.128281 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.128379 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zpnmm" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.128383 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.128920 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1cfc1793-db10-47a8-a152-0c86cc82be4c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-68c2v\" (UID: \"1cfc1793-db10-47a8-a152-0c86cc82be4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.129828 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.131627 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1366a10-f981-4c9a-aecd-415273b5b714-serving-cert\") pod \"console-operator-58897d9998-8dj62\" (UID: \"d1366a10-f981-4c9a-aecd-415273b5b714\") " pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.131941 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e62d3902-9b6c-48ec-af1f-4ee9823ccc12-apiservice-cert\") pod \"packageserver-d55dfcdfc-xtdpd\" (UID: \"e62d3902-9b6c-48ec-af1f-4ee9823ccc12\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.132097 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-serving-cert\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.135561 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.135606 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.136108 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7prrf"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137254 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-service-ca\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.119756 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-image-import-ca\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137257 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/961e712e-5760-4436-9be7-6bb6be2f8cf3-serving-cert\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137322 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49xvl\" (UniqueName: \"kubernetes.io/projected/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-kube-api-access-49xvl\") pod \"route-controller-manager-6576b87f9c-g92wn\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137351 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d039518e-be81-4d51-9798-7043a3b653e0-trusted-ca\") pod \"ingress-operator-5b745b69d9-58fw2\" (UID: \"d039518e-be81-4d51-9798-7043a3b653e0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137373 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137396 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fefa105a-53b5-45a0-ac11-f3078d78ae59-audit-dir\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137419 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-serving-cert\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137441 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137461 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137482 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137509 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzd9r\" (UniqueName: \"kubernetes.io/projected/b3217c27-f7c4-4b0f-9503-1968ce24eaba-kube-api-access-jzd9r\") pod \"openshift-controller-manager-operator-756b6f6bc6-rrb82\" (UID: \"b3217c27-f7c4-4b0f-9503-1968ce24eaba\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137534 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-audit\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137554 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137573 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-service-ca-bundle\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137593 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-config\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137612 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-serving-cert\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137632 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/961e712e-5760-4436-9be7-6bb6be2f8cf3-audit-dir\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137655 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-client-ca\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137676 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76c97829-08c2-4e93-ae1f-4788dca51fc4-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pxjp8\" (UID: \"76c97829-08c2-4e93-ae1f-4788dca51fc4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137697 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-serving-cert\") pod \"route-controller-manager-6576b87f9c-g92wn\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137720 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0ffcb835-7bd1-4cb3-b7cf-354595aa0b27-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-r8c2x\" (UID: \"0ffcb835-7bd1-4cb3-b7cf-354595aa0b27\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137742 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-config\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137760 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lph2s\" (UniqueName: \"kubernetes.io/projected/e0904a9e-81f3-4849-9f7d-90e2485a218a-kube-api-access-lph2s\") pod \"cluster-samples-operator-665b6dd947-rlhp2\" (UID: \"e0904a9e-81f3-4849-9f7d-90e2485a218a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137781 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm5wr\" (UniqueName: \"kubernetes.io/projected/9fd48e0d-81e5-49ed-8ced-4c382f96f831-kube-api-access-hm5wr\") pod \"dns-operator-744455d44c-jb4nn\" (UID: \"9fd48e0d-81e5-49ed-8ced-4c382f96f831\") " pod="openshift-dns-operator/dns-operator-744455d44c-jb4nn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137821 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c65812c5-b7ed-434b-bdef-bbbe26fe5c69-auth-proxy-config\") pod \"machine-approver-56656f9798-6zk5s\" (UID: \"c65812c5-b7ed-434b-bdef-bbbe26fe5c69\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137847 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-config\") pod \"route-controller-manager-6576b87f9c-g92wn\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137871 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-etcd-ca\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137893 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-etcd-client\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137914 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88b88dc4-a78f-489f-b970-f70828c68f6d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6n2dp\" (UID: \"88b88dc4-a78f-489f-b970-f70828c68f6d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137941 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3217c27-f7c4-4b0f-9503-1968ce24eaba-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rrb82\" (UID: \"b3217c27-f7c4-4b0f-9503-1968ce24eaba\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137961 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d039518e-be81-4d51-9798-7043a3b653e0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-58fw2\" (UID: \"d039518e-be81-4d51-9798-7043a3b653e0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.137978 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88b88dc4-a78f-489f-b970-f70828c68f6d-config\") pod \"kube-apiserver-operator-766d6c64bb-6n2dp\" (UID: \"88b88dc4-a78f-489f-b970-f70828c68f6d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.138009 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-config\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.138035 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c8cb8248-6e33-49aa-9422-f2e14c0e9015-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h8j9j\" (UID: \"c8cb8248-6e33-49aa-9422-f2e14c0e9015\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.138038 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-image-import-ca\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.138829 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.140341 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-config\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.140371 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-config\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.140435 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76c97829-08c2-4e93-ae1f-4788dca51fc4-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pxjp8\" (UID: \"76c97829-08c2-4e93-ae1f-4788dca51fc4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.140863 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.140940 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-etcd-serving-ca\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.141221 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c65812c5-b7ed-434b-bdef-bbbe26fe5c69-auth-proxy-config\") pod \"machine-approver-56656f9798-6zk5s\" (UID: \"c65812c5-b7ed-434b-bdef-bbbe26fe5c69\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.141443 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.141858 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/961e712e-5760-4436-9be7-6bb6be2f8cf3-audit-dir\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.142087 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-client-ca\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.142320 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-audit\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.142344 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-serving-cert\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.142394 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fefa105a-53b5-45a0-ac11-f3078d78ae59-audit-dir\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.142660 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-service-ca-bundle\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.142730 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.142784 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961e712e-5760-4436-9be7-6bb6be2f8cf3-config\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.143372 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-serving-cert\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.143904 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.143978 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-config\") pod \"route-controller-manager-6576b87f9c-g92wn\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.144714 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.144748 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/961e712e-5760-4436-9be7-6bb6be2f8cf3-etcd-client\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.145599 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.145648 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.147697 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pr9z4"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.148322 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-serving-cert\") pod \"route-controller-manager-6576b87f9c-g92wn\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.148519 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zpnmm"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.148904 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.149891 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.150078 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.151404 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.152955 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-449x6"] Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.153898 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-449x6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.170314 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.190499 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.209392 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.229996 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239024 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qqm8\" (UniqueName: \"kubernetes.io/projected/c8cb8248-6e33-49aa-9422-f2e14c0e9015-kube-api-access-9qqm8\") pod \"cluster-image-registry-operator-dc59b4c8b-h8j9j\" (UID: \"c8cb8248-6e33-49aa-9422-f2e14c0e9015\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239087 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8cb8248-6e33-49aa-9422-f2e14c0e9015-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h8j9j\" (UID: \"c8cb8248-6e33-49aa-9422-f2e14c0e9015\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239144 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0ffcb835-7bd1-4cb3-b7cf-354595aa0b27-proxy-tls\") pod \"machine-config-controller-84d6567774-r8c2x\" (UID: \"0ffcb835-7bd1-4cb3-b7cf-354595aa0b27\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239167 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9fd48e0d-81e5-49ed-8ced-4c382f96f831-metrics-tls\") pod \"dns-operator-744455d44c-jb4nn\" (UID: \"9fd48e0d-81e5-49ed-8ced-4c382f96f831\") " pod="openshift-dns-operator/dns-operator-744455d44c-jb4nn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239200 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f20b7d3-21f9-4a06-a160-df1bd47f389d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vcxd\" (UID: \"9f20b7d3-21f9-4a06-a160-df1bd47f389d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239225 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkkzw\" (UniqueName: \"kubernetes.io/projected/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-kube-api-access-pkkzw\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239248 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3217c27-f7c4-4b0f-9503-1968ce24eaba-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rrb82\" (UID: \"b3217c27-f7c4-4b0f-9503-1968ce24eaba\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239272 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz2v2\" (UniqueName: \"kubernetes.io/projected/19247691-b8da-4344-91e1-57bc628fca0a-kube-api-access-rz2v2\") pod \"migrator-59844c95c7-nfr9d\" (UID: \"19247691-b8da-4344-91e1-57bc628fca0a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nfr9d" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239307 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a833cc23-a6bc-4d07-95bf-d745791879d6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4wzs8\" (UID: \"a833cc23-a6bc-4d07-95bf-d745791879d6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239328 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c8cb8248-6e33-49aa-9422-f2e14c0e9015-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h8j9j\" (UID: \"c8cb8248-6e33-49aa-9422-f2e14c0e9015\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239363 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-config\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239385 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a833cc23-a6bc-4d07-95bf-d745791879d6-srv-cert\") pod \"olm-operator-6b444d44fb-4wzs8\" (UID: \"a833cc23-a6bc-4d07-95bf-d745791879d6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239418 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb4nc\" (UniqueName: \"kubernetes.io/projected/0ffcb835-7bd1-4cb3-b7cf-354595aa0b27-kube-api-access-jb4nc\") pod \"machine-config-controller-84d6567774-r8c2x\" (UID: \"0ffcb835-7bd1-4cb3-b7cf-354595aa0b27\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239496 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d039518e-be81-4d51-9798-7043a3b653e0-trusted-ca\") pod \"ingress-operator-5b745b69d9-58fw2\" (UID: \"d039518e-be81-4d51-9798-7043a3b653e0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239520 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzd9r\" (UniqueName: \"kubernetes.io/projected/b3217c27-f7c4-4b0f-9503-1968ce24eaba-kube-api-access-jzd9r\") pod \"openshift-controller-manager-operator-756b6f6bc6-rrb82\" (UID: \"b3217c27-f7c4-4b0f-9503-1968ce24eaba\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239548 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0ffcb835-7bd1-4cb3-b7cf-354595aa0b27-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-r8c2x\" (UID: \"0ffcb835-7bd1-4cb3-b7cf-354595aa0b27\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239592 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm5wr\" (UniqueName: \"kubernetes.io/projected/9fd48e0d-81e5-49ed-8ced-4c382f96f831-kube-api-access-hm5wr\") pod \"dns-operator-744455d44c-jb4nn\" (UID: \"9fd48e0d-81e5-49ed-8ced-4c382f96f831\") " pod="openshift-dns-operator/dns-operator-744455d44c-jb4nn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239616 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3217c27-f7c4-4b0f-9503-1968ce24eaba-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rrb82\" (UID: \"b3217c27-f7c4-4b0f-9503-1968ce24eaba\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239639 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-etcd-ca\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239660 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-etcd-client\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239681 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88b88dc4-a78f-489f-b970-f70828c68f6d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6n2dp\" (UID: \"88b88dc4-a78f-489f-b970-f70828c68f6d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239702 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d039518e-be81-4d51-9798-7043a3b653e0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-58fw2\" (UID: \"d039518e-be81-4d51-9798-7043a3b653e0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239722 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88b88dc4-a78f-489f-b970-f70828c68f6d-config\") pod \"kube-apiserver-operator-766d6c64bb-6n2dp\" (UID: \"88b88dc4-a78f-489f-b970-f70828c68f6d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239742 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c8cb8248-6e33-49aa-9422-f2e14c0e9015-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h8j9j\" (UID: \"c8cb8248-6e33-49aa-9422-f2e14c0e9015\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239774 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-serving-cert\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239796 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzvp8\" (UniqueName: \"kubernetes.io/projected/a833cc23-a6bc-4d07-95bf-d745791879d6-kube-api-access-kzvp8\") pod \"olm-operator-6b444d44fb-4wzs8\" (UID: \"a833cc23-a6bc-4d07-95bf-d745791879d6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239819 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d039518e-be81-4d51-9798-7043a3b653e0-metrics-tls\") pod \"ingress-operator-5b745b69d9-58fw2\" (UID: \"d039518e-be81-4d51-9798-7043a3b653e0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239847 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz4lx\" (UniqueName: \"kubernetes.io/projected/d039518e-be81-4d51-9798-7043a3b653e0-kube-api-access-wz4lx\") pod \"ingress-operator-5b745b69d9-58fw2\" (UID: \"d039518e-be81-4d51-9798-7043a3b653e0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239870 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88b88dc4-a78f-489f-b970-f70828c68f6d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6n2dp\" (UID: \"88b88dc4-a78f-489f-b970-f70828c68f6d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239892 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gzfw\" (UniqueName: \"kubernetes.io/projected/9f20b7d3-21f9-4a06-a160-df1bd47f389d-kube-api-access-6gzfw\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vcxd\" (UID: \"9f20b7d3-21f9-4a06-a160-df1bd47f389d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239917 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f20b7d3-21f9-4a06-a160-df1bd47f389d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vcxd\" (UID: \"9f20b7d3-21f9-4a06-a160-df1bd47f389d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.239942 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-etcd-service-ca\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.240679 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-etcd-service-ca\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.240999 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-config\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.241444 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-etcd-ca\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.241784 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0ffcb835-7bd1-4cb3-b7cf-354595aa0b27-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-r8c2x\" (UID: \"0ffcb835-7bd1-4cb3-b7cf-354595aa0b27\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.243406 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-serving-cert\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.243486 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-etcd-client\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.243617 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0ffcb835-7bd1-4cb3-b7cf-354595aa0b27-proxy-tls\") pod \"machine-config-controller-84d6567774-r8c2x\" (UID: \"0ffcb835-7bd1-4cb3-b7cf-354595aa0b27\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.244774 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.247113 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9fd48e0d-81e5-49ed-8ced-4c382f96f831-metrics-tls\") pod \"dns-operator-744455d44c-jb4nn\" (UID: \"9fd48e0d-81e5-49ed-8ced-4c382f96f831\") " pod="openshift-dns-operator/dns-operator-744455d44c-jb4nn" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.250492 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.269963 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.275013 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a833cc23-a6bc-4d07-95bf-d745791879d6-srv-cert\") pod \"olm-operator-6b444d44fb-4wzs8\" (UID: \"a833cc23-a6bc-4d07-95bf-d745791879d6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.295285 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.300301 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8cb8248-6e33-49aa-9422-f2e14c0e9015-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h8j9j\" (UID: \"c8cb8248-6e33-49aa-9422-f2e14c0e9015\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.310152 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.330342 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.350731 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.369775 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.373368 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a833cc23-a6bc-4d07-95bf-d745791879d6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4wzs8\" (UID: \"a833cc23-a6bc-4d07-95bf-d745791879d6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.391174 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.410444 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.430960 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.444799 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c8cb8248-6e33-49aa-9422-f2e14c0e9015-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h8j9j\" (UID: \"c8cb8248-6e33-49aa-9422-f2e14c0e9015\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.450281 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.470252 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.490933 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.510096 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.515168 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88b88dc4-a78f-489f-b970-f70828c68f6d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6n2dp\" (UID: \"88b88dc4-a78f-489f-b970-f70828c68f6d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.529982 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.549934 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.551708 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88b88dc4-a78f-489f-b970-f70828c68f6d-config\") pod \"kube-apiserver-operator-766d6c64bb-6n2dp\" (UID: \"88b88dc4-a78f-489f-b970-f70828c68f6d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.570903 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.590595 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.610366 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.614081 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f20b7d3-21f9-4a06-a160-df1bd47f389d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vcxd\" (UID: \"9f20b7d3-21f9-4a06-a160-df1bd47f389d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.630708 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.639960 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f20b7d3-21f9-4a06-a160-df1bd47f389d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vcxd\" (UID: \"9f20b7d3-21f9-4a06-a160-df1bd47f389d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.649695 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.669855 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.690781 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.710527 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.730394 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.770508 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.781286 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3217c27-f7c4-4b0f-9503-1968ce24eaba-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rrb82\" (UID: \"b3217c27-f7c4-4b0f-9503-1968ce24eaba\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.790915 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.810407 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.829779 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.833560 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3217c27-f7c4-4b0f-9503-1968ce24eaba-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rrb82\" (UID: \"b3217c27-f7c4-4b0f-9503-1968ce24eaba\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.850582 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.869567 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.890238 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.910777 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.914992 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d039518e-be81-4d51-9798-7043a3b653e0-metrics-tls\") pod \"ingress-operator-5b745b69d9-58fw2\" (UID: \"d039518e-be81-4d51-9798-7043a3b653e0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.936115 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.942422 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d039518e-be81-4d51-9798-7043a3b653e0-trusted-ca\") pod \"ingress-operator-5b745b69d9-58fw2\" (UID: \"d039518e-be81-4d51-9798-7043a3b653e0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.950049 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 02 09:39:14 crc kubenswrapper[4771]: I1002 09:39:14.990338 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.015101 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.029784 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.051034 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.068408 4771 request.go:700] Waited for 1.013906055s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/secrets?fieldSelector=metadata.name%3Dmarketplace-operator-dockercfg-5nsgg&limit=500&resourceVersion=0 Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.069621 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.090764 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.109144 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.130585 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.152051 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.170221 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.190054 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.210735 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.229744 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.250408 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.271779 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.290263 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.309759 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.330402 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.349780 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.370859 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.391037 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.410995 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.431147 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.451091 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.471239 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.491918 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.510745 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.531171 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.551653 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.570744 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.606785 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.611345 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.631764 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.651041 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.670726 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.690833 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.738227 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4qfx\" (UniqueName: \"kubernetes.io/projected/76c97829-08c2-4e93-ae1f-4788dca51fc4-kube-api-access-b4qfx\") pod \"openshift-apiserver-operator-796bbdcf4f-pxjp8\" (UID: \"76c97829-08c2-4e93-ae1f-4788dca51fc4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.744047 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvlgw\" (UniqueName: \"kubernetes.io/projected/c65812c5-b7ed-434b-bdef-bbbe26fe5c69-kube-api-access-jvlgw\") pod \"machine-approver-56656f9798-6zk5s\" (UID: \"c65812c5-b7ed-434b-bdef-bbbe26fe5c69\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.776815 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tntj5\" (UniqueName: \"kubernetes.io/projected/e3d49353-2942-437a-bfbb-2da3e5df246d-kube-api-access-tntj5\") pod \"oauth-openshift-558db77b4-gdzf9\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.785704 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4gln\" (UniqueName: \"kubernetes.io/projected/961e712e-5760-4436-9be7-6bb6be2f8cf3-kube-api-access-r4gln\") pod \"apiserver-76f77b778f-sv9ls\" (UID: \"961e712e-5760-4436-9be7-6bb6be2f8cf3\") " pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.805491 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjqfx\" (UniqueName: \"kubernetes.io/projected/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-kube-api-access-wjqfx\") pod \"console-f9d7485db-r664s\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.826503 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmhkn\" (UniqueName: \"kubernetes.io/projected/671d32ff-3e87-49bd-a4ec-60a495671813-kube-api-access-vmhkn\") pod \"openshift-config-operator-7777fb866f-zlrjz\" (UID: \"671d32ff-3e87-49bd-a4ec-60a495671813\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.832023 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.846646 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mcbd\" (UniqueName: \"kubernetes.io/projected/4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0-kube-api-access-8mcbd\") pod \"authentication-operator-69f744f599-nbmlz\" (UID: \"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.864792 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.869398 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f99ns\" (UniqueName: \"kubernetes.io/projected/1cfc1793-db10-47a8-a152-0c86cc82be4c-kube-api-access-f99ns\") pod \"machine-api-operator-5694c8668f-68c2v\" (UID: \"1cfc1793-db10-47a8-a152-0c86cc82be4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.877491 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.886520 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.889635 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-549tf\" (UniqueName: \"kubernetes.io/projected/e62d3902-9b6c-48ec-af1f-4ee9823ccc12-kube-api-access-549tf\") pod \"packageserver-d55dfcdfc-xtdpd\" (UID: \"e62d3902-9b6c-48ec-af1f-4ee9823ccc12\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.895615 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.903960 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.908179 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzxh7\" (UniqueName: \"kubernetes.io/projected/fefa105a-53b5-45a0-ac11-f3078d78ae59-kube-api-access-vzxh7\") pod \"apiserver-7bbb656c7d-rqp4t\" (UID: \"fefa105a-53b5-45a0-ac11-f3078d78ae59\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.931513 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz9tq\" (UniqueName: \"kubernetes.io/projected/77e60bc8-1d3d-4465-8dea-d7840a681b32-kube-api-access-gz9tq\") pod \"downloads-7954f5f757-vk4pw\" (UID: \"77e60bc8-1d3d-4465-8dea-d7840a681b32\") " pod="openshift-console/downloads-7954f5f757-vk4pw" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.948735 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtcgt\" (UniqueName: \"kubernetes.io/projected/d1366a10-f981-4c9a-aecd-415273b5b714-kube-api-access-mtcgt\") pod \"console-operator-58897d9998-8dj62\" (UID: \"d1366a10-f981-4c9a-aecd-415273b5b714\") " pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.972363 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.972841 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgmmt\" (UniqueName: \"kubernetes.io/projected/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-kube-api-access-bgmmt\") pod \"controller-manager-879f6c89f-p65v7\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.980463 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.987480 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" Oct 02 09:39:15 crc kubenswrapper[4771]: I1002 09:39:15.993645 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.007387 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.011469 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.032533 4771 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.043082 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.050247 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.056469 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.070241 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.089053 4771 request.go:700] Waited for 1.953275736s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Dcanary-serving-cert&limit=500&resourceVersion=0 Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.090977 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.109937 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vk4pw" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.126684 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.155081 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49xvl\" (UniqueName: \"kubernetes.io/projected/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-kube-api-access-49xvl\") pod \"route-controller-manager-6576b87f9c-g92wn\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.161530 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.169933 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.190889 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.241498 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qqm8\" (UniqueName: \"kubernetes.io/projected/c8cb8248-6e33-49aa-9422-f2e14c0e9015-kube-api-access-9qqm8\") pod \"cluster-image-registry-operator-dc59b4c8b-h8j9j\" (UID: \"c8cb8248-6e33-49aa-9422-f2e14c0e9015\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.258313 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lph2s\" (UniqueName: \"kubernetes.io/projected/e0904a9e-81f3-4849-9f7d-90e2485a218a-kube-api-access-lph2s\") pod \"cluster-samples-operator-665b6dd947-rlhp2\" (UID: \"e0904a9e-81f3-4849-9f7d-90e2485a218a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.270943 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkkzw\" (UniqueName: \"kubernetes.io/projected/4917c5c7-bed1-46fc-ab93-7f9959cc0b48-kube-api-access-pkkzw\") pod \"etcd-operator-b45778765-vz4s6\" (UID: \"4917c5c7-bed1-46fc-ab93-7f9959cc0b48\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.275387 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz2v2\" (UniqueName: \"kubernetes.io/projected/19247691-b8da-4344-91e1-57bc628fca0a-kube-api-access-rz2v2\") pod \"migrator-59844c95c7-nfr9d\" (UID: \"19247691-b8da-4344-91e1-57bc628fca0a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nfr9d" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.284767 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.289575 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzvp8\" (UniqueName: \"kubernetes.io/projected/a833cc23-a6bc-4d07-95bf-d745791879d6-kube-api-access-kzvp8\") pod \"olm-operator-6b444d44fb-4wzs8\" (UID: \"a833cc23-a6bc-4d07-95bf-d745791879d6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.293492 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.297600 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.305867 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c8cb8248-6e33-49aa-9422-f2e14c0e9015-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h8j9j\" (UID: \"c8cb8248-6e33-49aa-9422-f2e14c0e9015\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.311758 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nfr9d" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.314315 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.333788 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.336024 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d039518e-be81-4d51-9798-7043a3b653e0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-58fw2\" (UID: \"d039518e-be81-4d51-9798-7043a3b653e0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.355784 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb4nc\" (UniqueName: \"kubernetes.io/projected/0ffcb835-7bd1-4cb3-b7cf-354595aa0b27-kube-api-access-jb4nc\") pod \"machine-config-controller-84d6567774-r8c2x\" (UID: \"0ffcb835-7bd1-4cb3-b7cf-354595aa0b27\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.361461 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sv9ls"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.368410 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88b88dc4-a78f-489f-b970-f70828c68f6d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6n2dp\" (UID: \"88b88dc4-a78f-489f-b970-f70828c68f6d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.368617 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8dj62"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.386533 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gzfw\" (UniqueName: \"kubernetes.io/projected/9f20b7d3-21f9-4a06-a160-df1bd47f389d-kube-api-access-6gzfw\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vcxd\" (UID: \"9f20b7d3-21f9-4a06-a160-df1bd47f389d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.401376 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-r664s"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.403604 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.408892 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz4lx\" (UniqueName: \"kubernetes.io/projected/d039518e-be81-4d51-9798-7043a3b653e0-kube-api-access-wz4lx\") pod \"ingress-operator-5b745b69d9-58fw2\" (UID: \"d039518e-be81-4d51-9798-7043a3b653e0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.419186 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nbmlz"] Oct 02 09:39:16 crc kubenswrapper[4771]: W1002 09:39:16.424231 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod961e712e_5760_4436_9be7_6bb6be2f8cf3.slice/crio-10733017cfd5a97ae0b889d68d6e7ae9ca8b5d2856f5fa9ce56dcb753414f036 WatchSource:0}: Error finding container 10733017cfd5a97ae0b889d68d6e7ae9ca8b5d2856f5fa9ce56dcb753414f036: Status 404 returned error can't find the container with id 10733017cfd5a97ae0b889d68d6e7ae9ca8b5d2856f5fa9ce56dcb753414f036 Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.425015 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-gdzf9"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.435666 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm5wr\" (UniqueName: \"kubernetes.io/projected/9fd48e0d-81e5-49ed-8ced-4c382f96f831-kube-api-access-hm5wr\") pod \"dns-operator-744455d44c-jb4nn\" (UID: \"9fd48e0d-81e5-49ed-8ced-4c382f96f831\") " pod="openshift-dns-operator/dns-operator-744455d44c-jb4nn" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.437868 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" event={"ID":"e62d3902-9b6c-48ec-af1f-4ee9823ccc12","Type":"ContainerStarted","Data":"15d9c447639c685d181bf6e807e57436df6d42c49612002138b22c3dab775356"} Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.442065 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" event={"ID":"c65812c5-b7ed-434b-bdef-bbbe26fe5c69","Type":"ContainerStarted","Data":"7d4f49b2bb14c8335e75052c4bc83dcae88f1b105a65970227caad23494edbdc"} Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.448698 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-68c2v"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.449207 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.449347 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzd9r\" (UniqueName: \"kubernetes.io/projected/b3217c27-f7c4-4b0f-9503-1968ce24eaba-kube-api-access-jzd9r\") pod \"openshift-controller-manager-operator-756b6f6bc6-rrb82\" (UID: \"b3217c27-f7c4-4b0f-9503-1968ce24eaba\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.453753 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2" Oct 02 09:39:16 crc kubenswrapper[4771]: W1002 09:39:16.455052 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76c97829_08c2_4e93_ae1f_4788dca51fc4.slice/crio-db8ce4bf9fd6de457fd8b6ab53ac157346b4aafe7a82ccce82559b0769ca862e WatchSource:0}: Error finding container db8ce4bf9fd6de457fd8b6ab53ac157346b4aafe7a82ccce82559b0769ca862e: Status 404 returned error can't find the container with id db8ce4bf9fd6de457fd8b6ab53ac157346b4aafe7a82ccce82559b0769ca862e Oct 02 09:39:16 crc kubenswrapper[4771]: W1002 09:39:16.466230 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67d2ef8d_5a0c_48fa_8233_9bb23e794da0.slice/crio-ff4370f822102dba1a096f940eee0b085bf3681ee09b33d72af5987c86a68c36 WatchSource:0}: Error finding container ff4370f822102dba1a096f940eee0b085bf3681ee09b33d72af5987c86a68c36: Status 404 returned error can't find the container with id ff4370f822102dba1a096f940eee0b085bf3681ee09b33d72af5987c86a68c36 Oct 02 09:39:16 crc kubenswrapper[4771]: W1002 09:39:16.473064 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cfc1793_db10_47a8_a152_0c86cc82be4c.slice/crio-5d7c275c6da73c072e6d104aa4a0e179a40d23916046c1a4f57d746cdfb1cdff WatchSource:0}: Error finding container 5d7c275c6da73c072e6d104aa4a0e179a40d23916046c1a4f57d746cdfb1cdff: Status 404 returned error can't find the container with id 5d7c275c6da73c072e6d104aa4a0e179a40d23916046c1a4f57d746cdfb1cdff Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.474229 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.486975 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/675e6316-9501-4ac0-a134-de142b7bfcd4-registry-certificates\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.487053 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/675e6316-9501-4ac0-a134-de142b7bfcd4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.487104 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.487147 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d632e53b-2327-40eb-9609-2449f74541c0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j67pl\" (UID: \"d632e53b-2327-40eb-9609-2449f74541c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.487165 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d632e53b-2327-40eb-9609-2449f74541c0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j67pl\" (UID: \"d632e53b-2327-40eb-9609-2449f74541c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.487182 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46cb211b-6f4d-4484-8018-44c24c295287-config\") pod \"kube-controller-manager-operator-78b949d7b-pz4l5\" (UID: \"46cb211b-6f4d-4484-8018-44c24c295287\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.487228 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-bound-sa-token\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.487249 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-registry-tls\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.487266 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/675e6316-9501-4ac0-a134-de142b7bfcd4-trusted-ca\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.487284 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46cb211b-6f4d-4484-8018-44c24c295287-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pz4l5\" (UID: \"46cb211b-6f4d-4484-8018-44c24c295287\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.487300 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/46cb211b-6f4d-4484-8018-44c24c295287-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pz4l5\" (UID: \"46cb211b-6f4d-4484-8018-44c24c295287\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.487334 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/675e6316-9501-4ac0-a134-de142b7bfcd4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.487367 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d632e53b-2327-40eb-9609-2449f74541c0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j67pl\" (UID: \"d632e53b-2327-40eb-9609-2449f74541c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.487386 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbjm4\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-kube-api-access-sbjm4\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: E1002 09:39:16.487657 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:16.987640771 +0000 UTC m=+144.635325838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.510605 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.579965 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jb4nn" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.596193 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:16 crc kubenswrapper[4771]: E1002 09:39:16.596409 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.096365954 +0000 UTC m=+144.744051021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.599547 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9b94af46-45c2-4f5c-9a16-53a1cedbe976-certs\") pod \"machine-config-server-449x6\" (UID: \"9b94af46-45c2-4f5c-9a16-53a1cedbe976\") " pod="openshift-machine-config-operator/machine-config-server-449x6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.599645 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e09f2d35-5bde-444c-866d-1376247d4659-secret-volume\") pod \"collect-profiles-29323290-clj5r\" (UID: \"e09f2d35-5bde-444c-866d-1376247d4659\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.599703 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/61b3178b-0376-4db5-8369-af20d1a30056-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xbvnm\" (UID: \"61b3178b-0376-4db5-8369-af20d1a30056\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.599736 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/722c0898-dbc8-4f96-88f3-dcc328e61bd0-images\") pod \"machine-config-operator-74547568cd-wnvsk\" (UID: \"722c0898-dbc8-4f96-88f3-dcc328e61bd0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.599872 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjwkj\" (UniqueName: \"kubernetes.io/projected/61b3178b-0376-4db5-8369-af20d1a30056-kube-api-access-wjwkj\") pod \"control-plane-machine-set-operator-78cbb6b69f-xbvnm\" (UID: \"61b3178b-0376-4db5-8369-af20d1a30056\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.599911 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12fbfa54-08b0-41aa-ba11-80f84d2724cf-metrics-certs\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.599938 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-mountpoint-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600191 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-bound-sa-token\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600222 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/700cecb9-292a-4b4f-b244-528242ab782b-signing-key\") pod \"service-ca-9c57cc56f-l4x98\" (UID: \"700cecb9-292a-4b4f-b244-528242ab782b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600334 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-registry-tls\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600371 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/220faef5-5c96-4a65-9692-183a34d12a44-srv-cert\") pod \"catalog-operator-68c6474976-p2qn6\" (UID: \"220faef5-5c96-4a65-9692-183a34d12a44\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600404 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl7lk\" (UniqueName: \"kubernetes.io/projected/e3849bee-5c76-4308-8476-8abc1914d609-kube-api-access-fl7lk\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600476 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/675e6316-9501-4ac0-a134-de142b7bfcd4-trusted-ca\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600526 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-registration-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600652 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46cb211b-6f4d-4484-8018-44c24c295287-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pz4l5\" (UID: \"46cb211b-6f4d-4484-8018-44c24c295287\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600715 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7769bc8-aad5-4e10-99f0-968f5b3e8b57-config\") pod \"service-ca-operator-777779d784-z7kpd\" (UID: \"e7769bc8-aad5-4e10-99f0-968f5b3e8b57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600775 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/46cb211b-6f4d-4484-8018-44c24c295287-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pz4l5\" (UID: \"46cb211b-6f4d-4484-8018-44c24c295287\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600840 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfzd6\" (UniqueName: \"kubernetes.io/projected/378a97ab-3ce3-4000-b15c-61c92f834207-kube-api-access-wfzd6\") pod \"dns-default-n6z8c\" (UID: \"378a97ab-3ce3-4000-b15c-61c92f834207\") " pod="openshift-dns/dns-default-n6z8c" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600916 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/675e6316-9501-4ac0-a134-de142b7bfcd4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600949 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7p7c\" (UniqueName: \"kubernetes.io/projected/220faef5-5c96-4a65-9692-183a34d12a44-kube-api-access-p7p7c\") pod \"catalog-operator-68c6474976-p2qn6\" (UID: \"220faef5-5c96-4a65-9692-183a34d12a44\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.600983 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0524aac0-d101-4b59-b829-95390e99a159-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-q2xlm\" (UID: \"0524aac0-d101-4b59-b829-95390e99a159\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601077 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d632e53b-2327-40eb-9609-2449f74541c0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j67pl\" (UID: \"d632e53b-2327-40eb-9609-2449f74541c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601111 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/722c0898-dbc8-4f96-88f3-dcc328e61bd0-proxy-tls\") pod \"machine-config-operator-74547568cd-wnvsk\" (UID: \"722c0898-dbc8-4f96-88f3-dcc328e61bd0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601178 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12fbfa54-08b0-41aa-ba11-80f84d2724cf-service-ca-bundle\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601240 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbjm4\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-kube-api-access-sbjm4\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601335 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6x76\" (UniqueName: \"kubernetes.io/projected/700cecb9-292a-4b4f-b244-528242ab782b-kube-api-access-q6x76\") pod \"service-ca-9c57cc56f-l4x98\" (UID: \"700cecb9-292a-4b4f-b244-528242ab782b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601367 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvwrd\" (UniqueName: \"kubernetes.io/projected/e3fe7e38-10be-47d5-bd99-d01af7cbbf88-kube-api-access-jvwrd\") pod \"ingress-canary-zpnmm\" (UID: \"e3fe7e38-10be-47d5-bd99-d01af7cbbf88\") " pod="openshift-ingress-canary/ingress-canary-zpnmm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601438 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lbw9\" (UniqueName: \"kubernetes.io/projected/0524aac0-d101-4b59-b829-95390e99a159-kube-api-access-6lbw9\") pod \"package-server-manager-789f6589d5-q2xlm\" (UID: \"0524aac0-d101-4b59-b829-95390e99a159\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601492 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/675e6316-9501-4ac0-a134-de142b7bfcd4-registry-certificates\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601518 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a8ab108d-1536-402c-936a-b4689994b820-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7prrf\" (UID: \"a8ab108d-1536-402c-936a-b4689994b820\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7prrf" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601553 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw95d\" (UniqueName: \"kubernetes.io/projected/722c0898-dbc8-4f96-88f3-dcc328e61bd0-kube-api-access-cw95d\") pod \"machine-config-operator-74547568cd-wnvsk\" (UID: \"722c0898-dbc8-4f96-88f3-dcc328e61bd0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601606 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkzqb\" (UniqueName: \"kubernetes.io/projected/e7769bc8-aad5-4e10-99f0-968f5b3e8b57-kube-api-access-dkzqb\") pod \"service-ca-operator-777779d784-z7kpd\" (UID: \"e7769bc8-aad5-4e10-99f0-968f5b3e8b57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601637 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/12fbfa54-08b0-41aa-ba11-80f84d2724cf-default-certificate\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601668 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3fe7e38-10be-47d5-bd99-d01af7cbbf88-cert\") pod \"ingress-canary-zpnmm\" (UID: \"e3fe7e38-10be-47d5-bd99-d01af7cbbf88\") " pod="openshift-ingress-canary/ingress-canary-zpnmm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601730 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cx2j5\" (UID: \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601760 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpltl\" (UniqueName: \"kubernetes.io/projected/a8ab108d-1536-402c-936a-b4689994b820-kube-api-access-qpltl\") pod \"multus-admission-controller-857f4d67dd-7prrf\" (UID: \"a8ab108d-1536-402c-936a-b4689994b820\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7prrf" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601806 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-csi-data-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601850 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xqgk\" (UniqueName: \"kubernetes.io/projected/e09f2d35-5bde-444c-866d-1376247d4659-kube-api-access-6xqgk\") pod \"collect-profiles-29323290-clj5r\" (UID: \"e09f2d35-5bde-444c-866d-1376247d4659\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601880 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp5rh\" (UniqueName: \"kubernetes.io/projected/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-kube-api-access-bp5rh\") pod \"marketplace-operator-79b997595-cx2j5\" (UID: \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.601967 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/700cecb9-292a-4b4f-b244-528242ab782b-signing-cabundle\") pod \"service-ca-9c57cc56f-l4x98\" (UID: \"700cecb9-292a-4b4f-b244-528242ab782b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.602013 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/378a97ab-3ce3-4000-b15c-61c92f834207-metrics-tls\") pod \"dns-default-n6z8c\" (UID: \"378a97ab-3ce3-4000-b15c-61c92f834207\") " pod="openshift-dns/dns-default-n6z8c" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.602190 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/675e6316-9501-4ac0-a134-de142b7bfcd4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.602255 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e09f2d35-5bde-444c-866d-1376247d4659-config-volume\") pod \"collect-profiles-29323290-clj5r\" (UID: \"e09f2d35-5bde-444c-866d-1376247d4659\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.602318 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/722c0898-dbc8-4f96-88f3-dcc328e61bd0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-wnvsk\" (UID: \"722c0898-dbc8-4f96-88f3-dcc328e61bd0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603026 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r92dr\" (UniqueName: \"kubernetes.io/projected/12fbfa54-08b0-41aa-ba11-80f84d2724cf-kube-api-access-r92dr\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603070 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603180 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9b94af46-45c2-4f5c-9a16-53a1cedbe976-node-bootstrap-token\") pod \"machine-config-server-449x6\" (UID: \"9b94af46-45c2-4f5c-9a16-53a1cedbe976\") " pod="openshift-machine-config-operator/machine-config-server-449x6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603272 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-plugins-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603315 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/378a97ab-3ce3-4000-b15c-61c92f834207-config-volume\") pod \"dns-default-n6z8c\" (UID: \"378a97ab-3ce3-4000-b15c-61c92f834207\") " pod="openshift-dns/dns-default-n6z8c" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603395 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cx2j5\" (UID: \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603416 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7769bc8-aad5-4e10-99f0-968f5b3e8b57-serving-cert\") pod \"service-ca-operator-777779d784-z7kpd\" (UID: \"e7769bc8-aad5-4e10-99f0-968f5b3e8b57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603459 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-socket-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603747 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d632e53b-2327-40eb-9609-2449f74541c0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j67pl\" (UID: \"d632e53b-2327-40eb-9609-2449f74541c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603775 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/12fbfa54-08b0-41aa-ba11-80f84d2724cf-stats-auth\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603800 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d632e53b-2327-40eb-9609-2449f74541c0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j67pl\" (UID: \"d632e53b-2327-40eb-9609-2449f74541c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603826 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/220faef5-5c96-4a65-9692-183a34d12a44-profile-collector-cert\") pod \"catalog-operator-68c6474976-p2qn6\" (UID: \"220faef5-5c96-4a65-9692-183a34d12a44\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603865 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46cb211b-6f4d-4484-8018-44c24c295287-config\") pod \"kube-controller-manager-operator-78b949d7b-pz4l5\" (UID: \"46cb211b-6f4d-4484-8018-44c24c295287\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.603893 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm2mp\" (UniqueName: \"kubernetes.io/projected/9b94af46-45c2-4f5c-9a16-53a1cedbe976-kube-api-access-tm2mp\") pod \"machine-config-server-449x6\" (UID: \"9b94af46-45c2-4f5c-9a16-53a1cedbe976\") " pod="openshift-machine-config-operator/machine-config-server-449x6" Oct 02 09:39:16 crc kubenswrapper[4771]: E1002 09:39:16.605535 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.105507648 +0000 UTC m=+144.753192715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.605560 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/675e6316-9501-4ac0-a134-de142b7bfcd4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.607292 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d632e53b-2327-40eb-9609-2449f74541c0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j67pl\" (UID: \"d632e53b-2327-40eb-9609-2449f74541c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.607886 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46cb211b-6f4d-4484-8018-44c24c295287-config\") pod \"kube-controller-manager-operator-78b949d7b-pz4l5\" (UID: \"46cb211b-6f4d-4484-8018-44c24c295287\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.609939 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/675e6316-9501-4ac0-a134-de142b7bfcd4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.611561 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/675e6316-9501-4ac0-a134-de142b7bfcd4-registry-certificates\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.615313 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46cb211b-6f4d-4484-8018-44c24c295287-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pz4l5\" (UID: \"46cb211b-6f4d-4484-8018-44c24c295287\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.616250 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-registry-tls\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.617274 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d632e53b-2327-40eb-9609-2449f74541c0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j67pl\" (UID: \"d632e53b-2327-40eb-9609-2449f74541c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.623030 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/675e6316-9501-4ac0-a134-de142b7bfcd4-trusted-ca\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.631266 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.631570 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vk4pw"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.634250 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.640695 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.647738 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/46cb211b-6f4d-4484-8018-44c24c295287-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pz4l5\" (UID: \"46cb211b-6f4d-4484-8018-44c24c295287\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" Oct 02 09:39:16 crc kubenswrapper[4771]: W1002 09:39:16.661748 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfefa105a_53b5_45a0_ac11_f3078d78ae59.slice/crio-e1c65ca2af7b1c3e692aca9446df460f4db1b25988dedcd1dda11b9fa79d966f WatchSource:0}: Error finding container e1c65ca2af7b1c3e692aca9446df460f4db1b25988dedcd1dda11b9fa79d966f: Status 404 returned error can't find the container with id e1c65ca2af7b1c3e692aca9446df460f4db1b25988dedcd1dda11b9fa79d966f Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.668638 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbjm4\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-kube-api-access-sbjm4\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.682539 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-p65v7"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.683633 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.696375 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-bound-sa-token\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: W1002 09:39:16.701409 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77e60bc8_1d3d_4465_8dea_d7840a681b32.slice/crio-942a63f2c18da586f5ea9860ec4fde58d0ea86a3c644d9fe42c2a27257dbf35b WatchSource:0}: Error finding container 942a63f2c18da586f5ea9860ec4fde58d0ea86a3c644d9fe42c2a27257dbf35b: Status 404 returned error can't find the container with id 942a63f2c18da586f5ea9860ec4fde58d0ea86a3c644d9fe42c2a27257dbf35b Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.704888 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:16 crc kubenswrapper[4771]: E1002 09:39:16.705023 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.204999054 +0000 UTC m=+144.852684121 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705144 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xqgk\" (UniqueName: \"kubernetes.io/projected/e09f2d35-5bde-444c-866d-1376247d4659-kube-api-access-6xqgk\") pod \"collect-profiles-29323290-clj5r\" (UID: \"e09f2d35-5bde-444c-866d-1376247d4659\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705173 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp5rh\" (UniqueName: \"kubernetes.io/projected/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-kube-api-access-bp5rh\") pod \"marketplace-operator-79b997595-cx2j5\" (UID: \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705199 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/700cecb9-292a-4b4f-b244-528242ab782b-signing-cabundle\") pod \"service-ca-9c57cc56f-l4x98\" (UID: \"700cecb9-292a-4b4f-b244-528242ab782b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705220 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/378a97ab-3ce3-4000-b15c-61c92f834207-metrics-tls\") pod \"dns-default-n6z8c\" (UID: \"378a97ab-3ce3-4000-b15c-61c92f834207\") " pod="openshift-dns/dns-default-n6z8c" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705241 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e09f2d35-5bde-444c-866d-1376247d4659-config-volume\") pod \"collect-profiles-29323290-clj5r\" (UID: \"e09f2d35-5bde-444c-866d-1376247d4659\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705259 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/722c0898-dbc8-4f96-88f3-dcc328e61bd0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-wnvsk\" (UID: \"722c0898-dbc8-4f96-88f3-dcc328e61bd0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705287 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r92dr\" (UniqueName: \"kubernetes.io/projected/12fbfa54-08b0-41aa-ba11-80f84d2724cf-kube-api-access-r92dr\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705311 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705333 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9b94af46-45c2-4f5c-9a16-53a1cedbe976-node-bootstrap-token\") pod \"machine-config-server-449x6\" (UID: \"9b94af46-45c2-4f5c-9a16-53a1cedbe976\") " pod="openshift-machine-config-operator/machine-config-server-449x6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705350 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-plugins-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705366 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/378a97ab-3ce3-4000-b15c-61c92f834207-config-volume\") pod \"dns-default-n6z8c\" (UID: \"378a97ab-3ce3-4000-b15c-61c92f834207\") " pod="openshift-dns/dns-default-n6z8c" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705386 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cx2j5\" (UID: \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705402 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7769bc8-aad5-4e10-99f0-968f5b3e8b57-serving-cert\") pod \"service-ca-operator-777779d784-z7kpd\" (UID: \"e7769bc8-aad5-4e10-99f0-968f5b3e8b57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705425 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-socket-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705443 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/12fbfa54-08b0-41aa-ba11-80f84d2724cf-stats-auth\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705474 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/220faef5-5c96-4a65-9692-183a34d12a44-profile-collector-cert\") pod \"catalog-operator-68c6474976-p2qn6\" (UID: \"220faef5-5c96-4a65-9692-183a34d12a44\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705496 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm2mp\" (UniqueName: \"kubernetes.io/projected/9b94af46-45c2-4f5c-9a16-53a1cedbe976-kube-api-access-tm2mp\") pod \"machine-config-server-449x6\" (UID: \"9b94af46-45c2-4f5c-9a16-53a1cedbe976\") " pod="openshift-machine-config-operator/machine-config-server-449x6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705515 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9b94af46-45c2-4f5c-9a16-53a1cedbe976-certs\") pod \"machine-config-server-449x6\" (UID: \"9b94af46-45c2-4f5c-9a16-53a1cedbe976\") " pod="openshift-machine-config-operator/machine-config-server-449x6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705532 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e09f2d35-5bde-444c-866d-1376247d4659-secret-volume\") pod \"collect-profiles-29323290-clj5r\" (UID: \"e09f2d35-5bde-444c-866d-1376247d4659\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705551 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/61b3178b-0376-4db5-8369-af20d1a30056-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xbvnm\" (UID: \"61b3178b-0376-4db5-8369-af20d1a30056\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705574 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/722c0898-dbc8-4f96-88f3-dcc328e61bd0-images\") pod \"machine-config-operator-74547568cd-wnvsk\" (UID: \"722c0898-dbc8-4f96-88f3-dcc328e61bd0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705594 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjwkj\" (UniqueName: \"kubernetes.io/projected/61b3178b-0376-4db5-8369-af20d1a30056-kube-api-access-wjwkj\") pod \"control-plane-machine-set-operator-78cbb6b69f-xbvnm\" (UID: \"61b3178b-0376-4db5-8369-af20d1a30056\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705614 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12fbfa54-08b0-41aa-ba11-80f84d2724cf-metrics-certs\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705633 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-mountpoint-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705664 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/700cecb9-292a-4b4f-b244-528242ab782b-signing-key\") pod \"service-ca-9c57cc56f-l4x98\" (UID: \"700cecb9-292a-4b4f-b244-528242ab782b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" Oct 02 09:39:16 crc kubenswrapper[4771]: E1002 09:39:16.705671 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.205663533 +0000 UTC m=+144.853348590 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705708 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/220faef5-5c96-4a65-9692-183a34d12a44-srv-cert\") pod \"catalog-operator-68c6474976-p2qn6\" (UID: \"220faef5-5c96-4a65-9692-183a34d12a44\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705738 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl7lk\" (UniqueName: \"kubernetes.io/projected/e3849bee-5c76-4308-8476-8abc1914d609-kube-api-access-fl7lk\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705766 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-registration-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705795 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7769bc8-aad5-4e10-99f0-968f5b3e8b57-config\") pod \"service-ca-operator-777779d784-z7kpd\" (UID: \"e7769bc8-aad5-4e10-99f0-968f5b3e8b57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705815 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfzd6\" (UniqueName: \"kubernetes.io/projected/378a97ab-3ce3-4000-b15c-61c92f834207-kube-api-access-wfzd6\") pod \"dns-default-n6z8c\" (UID: \"378a97ab-3ce3-4000-b15c-61c92f834207\") " pod="openshift-dns/dns-default-n6z8c" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705835 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7p7c\" (UniqueName: \"kubernetes.io/projected/220faef5-5c96-4a65-9692-183a34d12a44-kube-api-access-p7p7c\") pod \"catalog-operator-68c6474976-p2qn6\" (UID: \"220faef5-5c96-4a65-9692-183a34d12a44\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705859 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0524aac0-d101-4b59-b829-95390e99a159-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-q2xlm\" (UID: \"0524aac0-d101-4b59-b829-95390e99a159\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705902 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/722c0898-dbc8-4f96-88f3-dcc328e61bd0-proxy-tls\") pod \"machine-config-operator-74547568cd-wnvsk\" (UID: \"722c0898-dbc8-4f96-88f3-dcc328e61bd0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705922 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12fbfa54-08b0-41aa-ba11-80f84d2724cf-service-ca-bundle\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705946 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6x76\" (UniqueName: \"kubernetes.io/projected/700cecb9-292a-4b4f-b244-528242ab782b-kube-api-access-q6x76\") pod \"service-ca-9c57cc56f-l4x98\" (UID: \"700cecb9-292a-4b4f-b244-528242ab782b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705964 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvwrd\" (UniqueName: \"kubernetes.io/projected/e3fe7e38-10be-47d5-bd99-d01af7cbbf88-kube-api-access-jvwrd\") pod \"ingress-canary-zpnmm\" (UID: \"e3fe7e38-10be-47d5-bd99-d01af7cbbf88\") " pod="openshift-ingress-canary/ingress-canary-zpnmm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705981 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lbw9\" (UniqueName: \"kubernetes.io/projected/0524aac0-d101-4b59-b829-95390e99a159-kube-api-access-6lbw9\") pod \"package-server-manager-789f6589d5-q2xlm\" (UID: \"0524aac0-d101-4b59-b829-95390e99a159\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.705998 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a8ab108d-1536-402c-936a-b4689994b820-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7prrf\" (UID: \"a8ab108d-1536-402c-936a-b4689994b820\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7prrf" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.706018 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw95d\" (UniqueName: \"kubernetes.io/projected/722c0898-dbc8-4f96-88f3-dcc328e61bd0-kube-api-access-cw95d\") pod \"machine-config-operator-74547568cd-wnvsk\" (UID: \"722c0898-dbc8-4f96-88f3-dcc328e61bd0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.706038 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkzqb\" (UniqueName: \"kubernetes.io/projected/e7769bc8-aad5-4e10-99f0-968f5b3e8b57-kube-api-access-dkzqb\") pod \"service-ca-operator-777779d784-z7kpd\" (UID: \"e7769bc8-aad5-4e10-99f0-968f5b3e8b57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.706057 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/12fbfa54-08b0-41aa-ba11-80f84d2724cf-default-certificate\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.706073 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3fe7e38-10be-47d5-bd99-d01af7cbbf88-cert\") pod \"ingress-canary-zpnmm\" (UID: \"e3fe7e38-10be-47d5-bd99-d01af7cbbf88\") " pod="openshift-ingress-canary/ingress-canary-zpnmm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.706095 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cx2j5\" (UID: \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.706110 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpltl\" (UniqueName: \"kubernetes.io/projected/a8ab108d-1536-402c-936a-b4689994b820-kube-api-access-qpltl\") pod \"multus-admission-controller-857f4d67dd-7prrf\" (UID: \"a8ab108d-1536-402c-936a-b4689994b820\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7prrf" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.706152 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-csi-data-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.706307 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-csi-data-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.707048 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/700cecb9-292a-4b4f-b244-528242ab782b-signing-cabundle\") pod \"service-ca-9c57cc56f-l4x98\" (UID: \"700cecb9-292a-4b4f-b244-528242ab782b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.713614 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-socket-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.714948 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cx2j5\" (UID: \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.715328 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-plugins-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.721068 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/722c0898-dbc8-4f96-88f3-dcc328e61bd0-images\") pod \"machine-config-operator-74547568cd-wnvsk\" (UID: \"722c0898-dbc8-4f96-88f3-dcc328e61bd0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.721593 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/722c0898-dbc8-4f96-88f3-dcc328e61bd0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-wnvsk\" (UID: \"722c0898-dbc8-4f96-88f3-dcc328e61bd0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.724199 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/220faef5-5c96-4a65-9692-183a34d12a44-srv-cert\") pod \"catalog-operator-68c6474976-p2qn6\" (UID: \"220faef5-5c96-4a65-9692-183a34d12a44\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.728626 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cx2j5\" (UID: \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.728733 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e09f2d35-5bde-444c-866d-1376247d4659-config-volume\") pod \"collect-profiles-29323290-clj5r\" (UID: \"e09f2d35-5bde-444c-866d-1376247d4659\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.728788 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-mountpoint-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.728842 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/378a97ab-3ce3-4000-b15c-61c92f834207-config-volume\") pod \"dns-default-n6z8c\" (UID: \"378a97ab-3ce3-4000-b15c-61c92f834207\") " pod="openshift-dns/dns-default-n6z8c" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.729277 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e3849bee-5c76-4308-8476-8abc1914d609-registration-dir\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.730390 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7769bc8-aad5-4e10-99f0-968f5b3e8b57-config\") pod \"service-ca-operator-777779d784-z7kpd\" (UID: \"e7769bc8-aad5-4e10-99f0-968f5b3e8b57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.730970 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/61b3178b-0376-4db5-8369-af20d1a30056-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xbvnm\" (UID: \"61b3178b-0376-4db5-8369-af20d1a30056\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.731609 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d632e53b-2327-40eb-9609-2449f74541c0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j67pl\" (UID: \"d632e53b-2327-40eb-9609-2449f74541c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.732495 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/722c0898-dbc8-4f96-88f3-dcc328e61bd0-proxy-tls\") pod \"machine-config-operator-74547568cd-wnvsk\" (UID: \"722c0898-dbc8-4f96-88f3-dcc328e61bd0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.747065 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/378a97ab-3ce3-4000-b15c-61c92f834207-metrics-tls\") pod \"dns-default-n6z8c\" (UID: \"378a97ab-3ce3-4000-b15c-61c92f834207\") " pod="openshift-dns/dns-default-n6z8c" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.747810 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e09f2d35-5bde-444c-866d-1376247d4659-secret-volume\") pod \"collect-profiles-29323290-clj5r\" (UID: \"e09f2d35-5bde-444c-866d-1376247d4659\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.747841 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/700cecb9-292a-4b4f-b244-528242ab782b-signing-key\") pod \"service-ca-9c57cc56f-l4x98\" (UID: \"700cecb9-292a-4b4f-b244-528242ab782b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.748745 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7769bc8-aad5-4e10-99f0-968f5b3e8b57-serving-cert\") pod \"service-ca-operator-777779d784-z7kpd\" (UID: \"e7769bc8-aad5-4e10-99f0-968f5b3e8b57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.748892 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a8ab108d-1536-402c-936a-b4689994b820-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7prrf\" (UID: \"a8ab108d-1536-402c-936a-b4689994b820\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7prrf" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.749881 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/220faef5-5c96-4a65-9692-183a34d12a44-profile-collector-cert\") pod \"catalog-operator-68c6474976-p2qn6\" (UID: \"220faef5-5c96-4a65-9692-183a34d12a44\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.750881 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3fe7e38-10be-47d5-bd99-d01af7cbbf88-cert\") pod \"ingress-canary-zpnmm\" (UID: \"e3fe7e38-10be-47d5-bd99-d01af7cbbf88\") " pod="openshift-ingress-canary/ingress-canary-zpnmm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.751634 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0524aac0-d101-4b59-b829-95390e99a159-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-q2xlm\" (UID: \"0524aac0-d101-4b59-b829-95390e99a159\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" Oct 02 09:39:16 crc kubenswrapper[4771]: W1002 09:39:16.766961 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8f91205_c6e3_4f73_ac1b_21a7e79620c1.slice/crio-620cea6fc4f81c9aa4602d0967a02256bdc0bdf3f69d0d3a84103a2b36958296 WatchSource:0}: Error finding container 620cea6fc4f81c9aa4602d0967a02256bdc0bdf3f69d0d3a84103a2b36958296: Status 404 returned error can't find the container with id 620cea6fc4f81c9aa4602d0967a02256bdc0bdf3f69d0d3a84103a2b36958296 Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.770155 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp5rh\" (UniqueName: \"kubernetes.io/projected/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-kube-api-access-bp5rh\") pod \"marketplace-operator-79b997595-cx2j5\" (UID: \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\") " pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.771187 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.783527 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12fbfa54-08b0-41aa-ba11-80f84d2724cf-service-ca-bundle\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.783554 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9b94af46-45c2-4f5c-9a16-53a1cedbe976-node-bootstrap-token\") pod \"machine-config-server-449x6\" (UID: \"9b94af46-45c2-4f5c-9a16-53a1cedbe976\") " pod="openshift-machine-config-operator/machine-config-server-449x6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.784648 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/12fbfa54-08b0-41aa-ba11-80f84d2724cf-default-certificate\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.784919 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xqgk\" (UniqueName: \"kubernetes.io/projected/e09f2d35-5bde-444c-866d-1376247d4659-kube-api-access-6xqgk\") pod \"collect-profiles-29323290-clj5r\" (UID: \"e09f2d35-5bde-444c-866d-1376247d4659\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.789281 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9b94af46-45c2-4f5c-9a16-53a1cedbe976-certs\") pod \"machine-config-server-449x6\" (UID: \"9b94af46-45c2-4f5c-9a16-53a1cedbe976\") " pod="openshift-machine-config-operator/machine-config-server-449x6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.790161 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/12fbfa54-08b0-41aa-ba11-80f84d2724cf-stats-auth\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.791156 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12fbfa54-08b0-41aa-ba11-80f84d2724cf-metrics-certs\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.806878 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:16 crc kubenswrapper[4771]: E1002 09:39:16.807213 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.307180827 +0000 UTC m=+144.954865894 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.807508 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: E1002 09:39:16.807771 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.307759484 +0000 UTC m=+144.955444551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.814670 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjwkj\" (UniqueName: \"kubernetes.io/projected/61b3178b-0376-4db5-8369-af20d1a30056-kube-api-access-wjwkj\") pod \"control-plane-machine-set-operator-78cbb6b69f-xbvnm\" (UID: \"61b3178b-0376-4db5-8369-af20d1a30056\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.829350 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6x76\" (UniqueName: \"kubernetes.io/projected/700cecb9-292a-4b4f-b244-528242ab782b-kube-api-access-q6x76\") pod \"service-ca-9c57cc56f-l4x98\" (UID: \"700cecb9-292a-4b4f-b244-528242ab782b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.854551 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvwrd\" (UniqueName: \"kubernetes.io/projected/e3fe7e38-10be-47d5-bd99-d01af7cbbf88-kube-api-access-jvwrd\") pod \"ingress-canary-zpnmm\" (UID: \"e3fe7e38-10be-47d5-bd99-d01af7cbbf88\") " pod="openshift-ingress-canary/ingress-canary-zpnmm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.867793 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lbw9\" (UniqueName: \"kubernetes.io/projected/0524aac0-d101-4b59-b829-95390e99a159-kube-api-access-6lbw9\") pod \"package-server-manager-789f6589d5-q2xlm\" (UID: \"0524aac0-d101-4b59-b829-95390e99a159\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.869589 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.883349 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r92dr\" (UniqueName: \"kubernetes.io/projected/12fbfa54-08b0-41aa-ba11-80f84d2724cf-kube-api-access-r92dr\") pod \"router-default-5444994796-z7f8f\" (UID: \"12fbfa54-08b0-41aa-ba11-80f84d2724cf\") " pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.887761 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkzqb\" (UniqueName: \"kubernetes.io/projected/e7769bc8-aad5-4e10-99f0-968f5b3e8b57-kube-api-access-dkzqb\") pod \"service-ca-operator-777779d784-z7kpd\" (UID: \"e7769bc8-aad5-4e10-99f0-968f5b3e8b57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" Oct 02 09:39:16 crc kubenswrapper[4771]: W1002 09:39:16.901650 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda833cc23_a6bc_4d07_95bf_d745791879d6.slice/crio-9f7047b823bff1da859abb069701fd7fcc4e552f32e9eb3260e6095ea19fb752 WatchSource:0}: Error finding container 9f7047b823bff1da859abb069701fd7fcc4e552f32e9eb3260e6095ea19fb752: Status 404 returned error can't find the container with id 9f7047b823bff1da859abb069701fd7fcc4e552f32e9eb3260e6095ea19fb752 Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.908281 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:16 crc kubenswrapper[4771]: E1002 09:39:16.908451 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.408423653 +0000 UTC m=+145.056108720 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.908885 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:16 crc kubenswrapper[4771]: E1002 09:39:16.909529 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.409515375 +0000 UTC m=+145.057200442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.920354 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw95d\" (UniqueName: \"kubernetes.io/projected/722c0898-dbc8-4f96-88f3-dcc328e61bd0-kube-api-access-cw95d\") pod \"machine-config-operator-74547568cd-wnvsk\" (UID: \"722c0898-dbc8-4f96-88f3-dcc328e61bd0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.929753 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpltl\" (UniqueName: \"kubernetes.io/projected/a8ab108d-1536-402c-936a-b4689994b820-kube-api-access-qpltl\") pod \"multus-admission-controller-857f4d67dd-7prrf\" (UID: \"a8ab108d-1536-402c-936a-b4689994b820\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7prrf" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.951941 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.954609 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vz4s6"] Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.960051 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm2mp\" (UniqueName: \"kubernetes.io/projected/9b94af46-45c2-4f5c-9a16-53a1cedbe976-kube-api-access-tm2mp\") pod \"machine-config-server-449x6\" (UID: \"9b94af46-45c2-4f5c-9a16-53a1cedbe976\") " pod="openshift-machine-config-operator/machine-config-server-449x6" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.968885 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfzd6\" (UniqueName: \"kubernetes.io/projected/378a97ab-3ce3-4000-b15c-61c92f834207-kube-api-access-wfzd6\") pod \"dns-default-n6z8c\" (UID: \"378a97ab-3ce3-4000-b15c-61c92f834207\") " pod="openshift-dns/dns-default-n6z8c" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.969347 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.989034 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl7lk\" (UniqueName: \"kubernetes.io/projected/e3849bee-5c76-4308-8476-8abc1914d609-kube-api-access-fl7lk\") pod \"csi-hostpathplugin-pr9z4\" (UID: \"e3849bee-5c76-4308-8476-8abc1914d609\") " pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:16 crc kubenswrapper[4771]: I1002 09:39:16.992415 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.007104 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.009311 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.009637 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7p7c\" (UniqueName: \"kubernetes.io/projected/220faef5-5c96-4a65-9692-183a34d12a44-kube-api-access-p7p7c\") pod \"catalog-operator-68c6474976-p2qn6\" (UID: \"220faef5-5c96-4a65-9692-183a34d12a44\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" Oct 02 09:39:17 crc kubenswrapper[4771]: E1002 09:39:17.014600 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.514576592 +0000 UTC m=+145.162261659 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.014921 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.021700 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7prrf" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.031111 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.037731 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.053781 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.056267 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.066256 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j"] Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.066472 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-n6z8c" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.073319 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" Oct 02 09:39:17 crc kubenswrapper[4771]: W1002 09:39:17.076994 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8cb8248_6e33_49aa_9422_f2e14c0e9015.slice/crio-30202908088eca7f10a9ccfac8a71e1c54c9c7663af206cb644c23a33268c087 WatchSource:0}: Error finding container 30202908088eca7f10a9ccfac8a71e1c54c9c7663af206cb644c23a33268c087: Status 404 returned error can't find the container with id 30202908088eca7f10a9ccfac8a71e1c54c9c7663af206cb644c23a33268c087 Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.081015 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zpnmm" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.111102 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:17 crc kubenswrapper[4771]: E1002 09:39:17.111470 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.611455822 +0000 UTC m=+145.259140889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.114638 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-nfr9d"] Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.120355 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.128678 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-449x6" Oct 02 09:39:17 crc kubenswrapper[4771]: W1002 09:39:17.139284 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd438f5d_c3ac_4a51_bc47_acd68fcc356a.slice/crio-4af5fcd51481b37ef71a69287f3a015d4b7226cf7fae0740a768d39dfb66524f WatchSource:0}: Error finding container 4af5fcd51481b37ef71a69287f3a015d4b7226cf7fae0740a768d39dfb66524f: Status 404 returned error can't find the container with id 4af5fcd51481b37ef71a69287f3a015d4b7226cf7fae0740a768d39dfb66524f Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.145036 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp"] Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.174998 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2"] Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.211777 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:17 crc kubenswrapper[4771]: E1002 09:39:17.212075 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.712046949 +0000 UTC m=+145.359732156 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.212152 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:17 crc kubenswrapper[4771]: E1002 09:39:17.212524 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.712508422 +0000 UTC m=+145.360193669 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:17 crc kubenswrapper[4771]: W1002 09:39:17.262830 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19247691_b8da_4344_91e1_57bc628fca0a.slice/crio-30838f905879225442a33b920ccd235eb041ba13b7c205f6f51dfac65db6fc09 WatchSource:0}: Error finding container 30838f905879225442a33b920ccd235eb041ba13b7c205f6f51dfac65db6fc09: Status 404 returned error can't find the container with id 30838f905879225442a33b920ccd235eb041ba13b7c205f6f51dfac65db6fc09 Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.270357 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd"] Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.300332 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.319116 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:17 crc kubenswrapper[4771]: E1002 09:39:17.323104 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.823076118 +0000 UTC m=+145.470761185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.388149 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jb4nn"] Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.405838 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82"] Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.424091 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:17 crc kubenswrapper[4771]: E1002 09:39:17.424598 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:17.924583462 +0000 UTC m=+145.572268529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.484770 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x"] Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.493638 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" event={"ID":"88b88dc4-a78f-489f-b970-f70828c68f6d","Type":"ContainerStarted","Data":"ec8fdf3979ec1df0fd26a7ce0d8210e7c63db009fded05f031343e54d48816ea"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.500699 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r664s" event={"ID":"67d2ef8d-5a0c-48fa-8233-9bb23e794da0","Type":"ContainerStarted","Data":"ff4370f822102dba1a096f940eee0b085bf3681ee09b33d72af5987c86a68c36"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.510940 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" event={"ID":"c65812c5-b7ed-434b-bdef-bbbe26fe5c69","Type":"ContainerStarted","Data":"2001858f350eae9562cbcc146de59aa01f079be5f26ad779f646c06f26472a85"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.528793 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:17 crc kubenswrapper[4771]: E1002 09:39:17.529349 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:18.029329019 +0000 UTC m=+145.677014086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.534944 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" event={"ID":"671d32ff-3e87-49bd-a4ec-60a495671813","Type":"ContainerStarted","Data":"42598fbd0196d472eebd6fdb2cbf61ced884ac4c3b93f77e1de4ba04f2c5a198"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.539604 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" event={"ID":"e3d49353-2942-437a-bfbb-2da3e5df246d","Type":"ContainerStarted","Data":"0057b189ebe55c1ddfe4c29b2de981b955ab0afac5a894c5fabbe757cecd9762"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.544146 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" event={"ID":"fefa105a-53b5-45a0-ac11-f3078d78ae59","Type":"ContainerStarted","Data":"e1c65ca2af7b1c3e692aca9446df460f4db1b25988dedcd1dda11b9fa79d966f"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.546816 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" event={"ID":"a833cc23-a6bc-4d07-95bf-d745791879d6","Type":"ContainerStarted","Data":"9f7047b823bff1da859abb069701fd7fcc4e552f32e9eb3260e6095ea19fb752"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.547031 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5"] Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.549613 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" event={"ID":"1cfc1793-db10-47a8-a152-0c86cc82be4c","Type":"ContainerStarted","Data":"3b67e03b8615cd97eb744cac65e34bd4ecc5046d9fbfa8c7e9828c2aeb0c6881"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.549674 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" event={"ID":"1cfc1793-db10-47a8-a152-0c86cc82be4c","Type":"ContainerStarted","Data":"5d7c275c6da73c072e6d104aa4a0e179a40d23916046c1a4f57d746cdfb1cdff"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.552611 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" event={"ID":"c8cb8248-6e33-49aa-9422-f2e14c0e9015","Type":"ContainerStarted","Data":"30202908088eca7f10a9ccfac8a71e1c54c9c7663af206cb644c23a33268c087"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.553754 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" event={"ID":"4917c5c7-bed1-46fc-ab93-7f9959cc0b48","Type":"ContainerStarted","Data":"fc5963ebffdc02a6761d61f8c2e9fd9c39d0d8219b6c3999eee0ddb028db1829"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.558539 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8dj62" event={"ID":"d1366a10-f981-4c9a-aecd-415273b5b714","Type":"ContainerStarted","Data":"4a8e2c92743a77245f5f8627197f3d2713423c687d8339efb0dcbce6cd9c7f31"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.559790 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" event={"ID":"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0","Type":"ContainerStarted","Data":"e336f2765ab351effdeab3ccf581c05aca3bd5d31a0052322f71c6ba6ea03468"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.564114 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" event={"ID":"b8f91205-c6e3-4f73-ac1b-21a7e79620c1","Type":"ContainerStarted","Data":"620cea6fc4f81c9aa4602d0967a02256bdc0bdf3f69d0d3a84103a2b36958296"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.565184 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nfr9d" event={"ID":"19247691-b8da-4344-91e1-57bc628fca0a","Type":"ContainerStarted","Data":"30838f905879225442a33b920ccd235eb041ba13b7c205f6f51dfac65db6fc09"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.569445 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" event={"ID":"dd438f5d-c3ac-4a51-bc47-acd68fcc356a","Type":"ContainerStarted","Data":"4af5fcd51481b37ef71a69287f3a015d4b7226cf7fae0740a768d39dfb66524f"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.571368 4771 generic.go:334] "Generic (PLEG): container finished" podID="961e712e-5760-4436-9be7-6bb6be2f8cf3" containerID="0fc978a50883d06eb0c796535fc68a13f5d37b1d6b10daa3812c3b7d50bc6153" exitCode=0 Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.571452 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" event={"ID":"961e712e-5760-4436-9be7-6bb6be2f8cf3","Type":"ContainerDied","Data":"0fc978a50883d06eb0c796535fc68a13f5d37b1d6b10daa3812c3b7d50bc6153"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.571522 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" event={"ID":"961e712e-5760-4436-9be7-6bb6be2f8cf3","Type":"ContainerStarted","Data":"10733017cfd5a97ae0b889d68d6e7ae9ca8b5d2856f5fa9ce56dcb753414f036"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.572984 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" event={"ID":"e62d3902-9b6c-48ec-af1f-4ee9823ccc12","Type":"ContainerStarted","Data":"444678c126d83c8854030ba63863e619976d4ead3893fe3492ee180a6afdc748"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.573135 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.574971 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vk4pw" event={"ID":"77e60bc8-1d3d-4465-8dea-d7840a681b32","Type":"ContainerStarted","Data":"942a63f2c18da586f5ea9860ec4fde58d0ea86a3c644d9fe42c2a27257dbf35b"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.578331 4771 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-xtdpd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.14:5443/healthz\": dial tcp 10.217.0.14:5443: connect: connection refused" start-of-body= Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.578392 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" podUID="e62d3902-9b6c-48ec-af1f-4ee9823ccc12" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.14:5443/healthz\": dial tcp 10.217.0.14:5443: connect: connection refused" Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.580561 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" event={"ID":"76c97829-08c2-4e93-ae1f-4788dca51fc4","Type":"ContainerStarted","Data":"db8ce4bf9fd6de457fd8b6ab53ac157346b4aafe7a82ccce82559b0769ca862e"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.592750 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" event={"ID":"9f20b7d3-21f9-4a06-a160-df1bd47f389d","Type":"ContainerStarted","Data":"22b768ae2c2cbc660c9220741a3e4c8b6b169168ad1406b197bc7b609e1c6273"} Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.606926 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2"] Oct 02 09:39:17 crc kubenswrapper[4771]: W1002 09:39:17.625426 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46cb211b_6f4d_4484_8018_44c24c295287.slice/crio-4a47bb4fa5ed564d66c7485105bf52e851247f562b05c2ed10669c4ff57cb0de WatchSource:0}: Error finding container 4a47bb4fa5ed564d66c7485105bf52e851247f562b05c2ed10669c4ff57cb0de: Status 404 returned error can't find the container with id 4a47bb4fa5ed564d66c7485105bf52e851247f562b05c2ed10669c4ff57cb0de Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.630927 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:17 crc kubenswrapper[4771]: E1002 09:39:17.632564 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:18.132544412 +0000 UTC m=+145.780229659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:17 crc kubenswrapper[4771]: W1002 09:39:17.640820 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd039518e_be81_4d51_9798_7043a3b653e0.slice/crio-c5ca0a918a34120004187d1dcb9fd1a2a554c28e59dd007cdbd13de31ba2f42b WatchSource:0}: Error finding container c5ca0a918a34120004187d1dcb9fd1a2a554c28e59dd007cdbd13de31ba2f42b: Status 404 returned error can't find the container with id c5ca0a918a34120004187d1dcb9fd1a2a554c28e59dd007cdbd13de31ba2f42b Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.732409 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:17 crc kubenswrapper[4771]: E1002 09:39:17.732827 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:18.23281011 +0000 UTC m=+145.880495177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.872807 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:17 crc kubenswrapper[4771]: E1002 09:39:17.873739 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:18.373724833 +0000 UTC m=+146.021409900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.890459 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl"] Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.903039 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cx2j5"] Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.939474 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r"] Oct 02 09:39:17 crc kubenswrapper[4771]: I1002 09:39:17.977492 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:17 crc kubenswrapper[4771]: E1002 09:39:17.977916 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:18.477893754 +0000 UTC m=+146.125578821 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.080578 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:18 crc kubenswrapper[4771]: E1002 09:39:18.081275 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:18.581252461 +0000 UTC m=+146.228937698 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.183244 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:18 crc kubenswrapper[4771]: E1002 09:39:18.183861 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:18.683840436 +0000 UTC m=+146.331525503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.219432 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk"] Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.225948 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm"] Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.231413 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zpnmm"] Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.234454 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd"] Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.285059 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.286201 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-n6z8c"] Oct 02 09:39:18 crc kubenswrapper[4771]: E1002 09:39:18.287822 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:18.787797781 +0000 UTC m=+146.435482848 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.392979 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:18 crc kubenswrapper[4771]: E1002 09:39:18.393918 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:18.893896147 +0000 UTC m=+146.541581214 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.490877 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pr9z4"] Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.495756 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:18 crc kubenswrapper[4771]: E1002 09:39:18.496161 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:18.996147883 +0000 UTC m=+146.643832950 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.510298 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" podStartSLOduration=120.510269471 podStartE2EDuration="2m0.510269471s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:18.504054121 +0000 UTC m=+146.151739188" watchObservedRunningTime="2025-10-02 09:39:18.510269471 +0000 UTC m=+146.157954538" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.513419 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7prrf"] Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.537278 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm"] Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.599326 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:18 crc kubenswrapper[4771]: E1002 09:39:18.599609 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:19.099551621 +0000 UTC m=+146.747236688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.600052 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:18 crc kubenswrapper[4771]: E1002 09:39:18.600553 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:19.10053448 +0000 UTC m=+146.748219547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.605076 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-l4x98"] Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.610807 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" event={"ID":"e7769bc8-aad5-4e10-99f0-968f5b3e8b57","Type":"ContainerStarted","Data":"a00af520534921be6bc781c739f3de734a08960bd66e879e2721d7063a2e3a0e"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.617048 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" event={"ID":"4bdc3d24-cf7e-447e-b6ed-63cd10b8a7f0","Type":"ContainerStarted","Data":"9d44b52a5abbdb6875f713934fb9b24ba34d0c076ec7cc2124670a767ac0da31"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.623439 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6"] Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.632379 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm" event={"ID":"61b3178b-0376-4db5-8369-af20d1a30056","Type":"ContainerStarted","Data":"365ec70a784745d2ee6674d07a683ae741e670a0fab77414e3ca909e1e09820a"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.638629 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-nbmlz" podStartSLOduration=121.63861176 podStartE2EDuration="2m1.63861176s" podCreationTimestamp="2025-10-02 09:37:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:18.638458466 +0000 UTC m=+146.286143553" watchObservedRunningTime="2025-10-02 09:39:18.63861176 +0000 UTC m=+146.286296847" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.643008 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zpnmm" event={"ID":"e3fe7e38-10be-47d5-bd99-d01af7cbbf88","Type":"ContainerStarted","Data":"844afae0ff6e27208f226c8b932446b6faa4692a039343fade5cb2dcccf15c1e"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.648534 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2" event={"ID":"e0904a9e-81f3-4849-9f7d-90e2485a218a","Type":"ContainerStarted","Data":"90a94901e807e288d1972b9558fcb1f7f16964487bdfd0cd9b3bba348e5a9e60"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.656164 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8dj62" event={"ID":"d1366a10-f981-4c9a-aecd-415273b5b714","Type":"ContainerStarted","Data":"40d73c9b002384a6b96655e4ea98da4d9828edb0016f099da57199c633e3c08b"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.656447 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.660396 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-449x6" event={"ID":"9b94af46-45c2-4f5c-9a16-53a1cedbe976","Type":"ContainerStarted","Data":"0e491b2f1633792330f0a91720d06290c26f488858d593bd57a7c2d98e3cf1bb"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.675629 4771 patch_prober.go:28] interesting pod/console-operator-58897d9998-8dj62 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.675742 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-8dj62" podUID="d1366a10-f981-4c9a-aecd-415273b5b714" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.678093 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-8dj62" podStartSLOduration=120.678076081 podStartE2EDuration="2m0.678076081s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:18.677576877 +0000 UTC m=+146.325261944" watchObservedRunningTime="2025-10-02 09:39:18.678076081 +0000 UTC m=+146.325761158" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.680915 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" event={"ID":"e3d49353-2942-437a-bfbb-2da3e5df246d","Type":"ContainerStarted","Data":"07dfd1351c030d8adeae28d10382941b51ee980774e191ae260cc34793f5f7d7"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.681112 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.688390 4771 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-gdzf9 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" start-of-body= Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.688498 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" podUID="e3d49353-2942-437a-bfbb-2da3e5df246d" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.693093 4771 generic.go:334] "Generic (PLEG): container finished" podID="671d32ff-3e87-49bd-a4ec-60a495671813" containerID="6fb949e1f8ba16d1be1a9c809a8c45646b029870970b594f83fa93b78848ab3b" exitCode=0 Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.693195 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" event={"ID":"671d32ff-3e87-49bd-a4ec-60a495671813","Type":"ContainerDied","Data":"6fb949e1f8ba16d1be1a9c809a8c45646b029870970b594f83fa93b78848ab3b"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.694780 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" event={"ID":"d039518e-be81-4d51-9798-7043a3b653e0","Type":"ContainerStarted","Data":"c5ca0a918a34120004187d1dcb9fd1a2a554c28e59dd007cdbd13de31ba2f42b"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.697149 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-z7f8f" event={"ID":"12fbfa54-08b0-41aa-ba11-80f84d2724cf","Type":"ContainerStarted","Data":"55ae0a5f74eed6d3b79e1eede02ed3a8b02ebce578fcc4e406c742b86b4c2903"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.698596 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" event={"ID":"b3217c27-f7c4-4b0f-9503-1968ce24eaba","Type":"ContainerStarted","Data":"c9dc00cd7a3f2231cce5af27f193fb85d04adebc0de34808ea5180f7ac9f3783"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.701442 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:18 crc kubenswrapper[4771]: E1002 09:39:18.701804 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:19.201768066 +0000 UTC m=+146.849453303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.706160 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" event={"ID":"e09f2d35-5bde-444c-866d-1376247d4659","Type":"ContainerStarted","Data":"a6cba2732c1f6e1e34d54df58abcebab9fb016f36dab507d6da83f6f9612a29c"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.708039 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" podStartSLOduration=121.708026717 podStartE2EDuration="2m1.708026717s" podCreationTimestamp="2025-10-02 09:37:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:18.706928005 +0000 UTC m=+146.354613072" watchObservedRunningTime="2025-10-02 09:39:18.708026717 +0000 UTC m=+146.355711784" Oct 02 09:39:18 crc kubenswrapper[4771]: W1002 09:39:18.717044 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod700cecb9_292a_4b4f_b244_528242ab782b.slice/crio-1584ea148af6fd36f3c3bb7fc16f30430db50a46534abff3e7b877c29c2c5a41 WatchSource:0}: Error finding container 1584ea148af6fd36f3c3bb7fc16f30430db50a46534abff3e7b877c29c2c5a41: Status 404 returned error can't find the container with id 1584ea148af6fd36f3c3bb7fc16f30430db50a46534abff3e7b877c29c2c5a41 Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.717378 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r664s" event={"ID":"67d2ef8d-5a0c-48fa-8233-9bb23e794da0","Type":"ContainerStarted","Data":"a118fa7267dd547ef56f471f662395be2a9720da4117eb6fef6014523b8ff519"} Oct 02 09:39:18 crc kubenswrapper[4771]: W1002 09:39:18.721630 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod220faef5_5c96_4a65_9692_183a34d12a44.slice/crio-aa7829c4aab2c688717f729c6563002a5c31280b1808dc194a2950a61db91d27 WatchSource:0}: Error finding container aa7829c4aab2c688717f729c6563002a5c31280b1808dc194a2950a61db91d27: Status 404 returned error can't find the container with id aa7829c4aab2c688717f729c6563002a5c31280b1808dc194a2950a61db91d27 Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.721656 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" event={"ID":"b8f91205-c6e3-4f73-ac1b-21a7e79620c1","Type":"ContainerStarted","Data":"ac526b310389f32d0d8b4b4cd84b4e0fb6a37b70c497f1a4d1f96ab23e183146"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.722250 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.724031 4771 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-p65v7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.724072 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" podUID="b8f91205-c6e3-4f73-ac1b-21a7e79620c1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.729717 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-n6z8c" event={"ID":"378a97ab-3ce3-4000-b15c-61c92f834207","Type":"ContainerStarted","Data":"bab4936495fde6d81e74c75e0407dc333d63aa1328b6d3f19f8cb49d982f836e"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.754967 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" event={"ID":"76c97829-08c2-4e93-ae1f-4788dca51fc4","Type":"ContainerStarted","Data":"4a4e66bd5c59e5af8cd8ed8b6672de0c5fdf3743a5ee378d78df34f695b847de"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.763619 4771 generic.go:334] "Generic (PLEG): container finished" podID="fefa105a-53b5-45a0-ac11-f3078d78ae59" containerID="a8af287f6a8a86ca20c804a90d34ba664160eb7d21562c21d94e91d44333032f" exitCode=0 Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.763718 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" event={"ID":"fefa105a-53b5-45a0-ac11-f3078d78ae59","Type":"ContainerDied","Data":"a8af287f6a8a86ca20c804a90d34ba664160eb7d21562c21d94e91d44333032f"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.771971 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jb4nn" event={"ID":"9fd48e0d-81e5-49ed-8ced-4c382f96f831","Type":"ContainerStarted","Data":"137ded8eb633fd8aae342e1122c7b823588cc1b8da69dcadc28898683fcc51a8"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.786736 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" event={"ID":"0ffcb835-7bd1-4cb3-b7cf-354595aa0b27","Type":"ContainerStarted","Data":"1da1ac11a3e8b38cef902929c9749a738005af3c76a3091038751ff61c64a84b"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.788743 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" podStartSLOduration=120.788722789 podStartE2EDuration="2m0.788722789s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:18.753243304 +0000 UTC m=+146.400928391" watchObservedRunningTime="2025-10-02 09:39:18.788722789 +0000 UTC m=+146.436407856" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.806884 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:18 crc kubenswrapper[4771]: E1002 09:39:18.813811 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:19.313778123 +0000 UTC m=+146.961463190 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.824770 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" event={"ID":"46cb211b-6f4d-4484-8018-44c24c295287","Type":"ContainerStarted","Data":"4a47bb4fa5ed564d66c7485105bf52e851247f562b05c2ed10669c4ff57cb0de"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.829789 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" event={"ID":"9f20b7d3-21f9-4a06-a160-df1bd47f389d","Type":"ContainerStarted","Data":"61be148db7937645bdc3f48d70cbc7964c52589bef2a8dba177018e2df2fd94b"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.840293 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-r664s" podStartSLOduration=120.840273829 podStartE2EDuration="2m0.840273829s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:18.789509492 +0000 UTC m=+146.437194569" watchObservedRunningTime="2025-10-02 09:39:18.840273829 +0000 UTC m=+146.487958896" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.847180 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vk4pw" event={"ID":"77e60bc8-1d3d-4465-8dea-d7840a681b32","Type":"ContainerStarted","Data":"79769cf9608f05c46ac87c4ce509d61f44810c565b8d8e1dd6d064e2c4ea7008"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.847693 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-vk4pw" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.849544 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.849618 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.854491 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" event={"ID":"5c65c130-1f41-42f8-be1a-9c8f17bf16dc","Type":"ContainerStarted","Data":"03d11622d647c0499a85d8d53e22db1afe6bd35bc6a9a3f94e598b6c44103764"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.864528 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" event={"ID":"722c0898-dbc8-4f96-88f3-dcc328e61bd0","Type":"ContainerStarted","Data":"2ad90e46a3196f75a7756df419f924e663f75489f9a24d18d0d6be6ff467c9ee"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.908325 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:18 crc kubenswrapper[4771]: E1002 09:39:18.909276 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:19.409248063 +0000 UTC m=+147.056933130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.912054 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxjp8" podStartSLOduration=121.912043173 podStartE2EDuration="2m1.912043173s" podCreationTimestamp="2025-10-02 09:37:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:18.868215367 +0000 UTC m=+146.515900434" watchObservedRunningTime="2025-10-02 09:39:18.912043173 +0000 UTC m=+146.559728240" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.913968 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-vk4pw" podStartSLOduration=120.913960509 podStartE2EDuration="2m0.913960509s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:18.913719112 +0000 UTC m=+146.561404179" watchObservedRunningTime="2025-10-02 09:39:18.913960509 +0000 UTC m=+146.561645576" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.916971 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.922084 4771 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-g92wn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.922146 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" podUID="dd438f5d-c3ac-4a51-bc47-acd68fcc356a" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.953571 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" event={"ID":"a833cc23-a6bc-4d07-95bf-d745791879d6","Type":"ContainerStarted","Data":"df3e509edbfdf2304c36dfc49b3687e78295df3661a90c4cd2559423f5c49e6b"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.954249 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.961475 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" event={"ID":"d632e53b-2327-40eb-9609-2449f74541c0","Type":"ContainerStarted","Data":"110c062917606cee2e1201ebd109f2c9606eef9a068f6a10d171459cc890a294"} Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.961585 4771 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-xtdpd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.14:5443/healthz\": dial tcp 10.217.0.14:5443: connect: connection refused" start-of-body= Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.961623 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" podUID="e62d3902-9b6c-48ec-af1f-4ee9823ccc12" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.14:5443/healthz\": dial tcp 10.217.0.14:5443: connect: connection refused" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.963694 4771 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-4wzs8 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.963729 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" podUID="a833cc23-a6bc-4d07-95bf-d745791879d6" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.981439 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" podStartSLOduration=120.981421149 podStartE2EDuration="2m0.981421149s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:18.98076135 +0000 UTC m=+146.628446417" watchObservedRunningTime="2025-10-02 09:39:18.981421149 +0000 UTC m=+146.629106216" Oct 02 09:39:18 crc kubenswrapper[4771]: I1002 09:39:18.983657 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vcxd" podStartSLOduration=120.983648783 podStartE2EDuration="2m0.983648783s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:18.946844189 +0000 UTC m=+146.594529256" watchObservedRunningTime="2025-10-02 09:39:18.983648783 +0000 UTC m=+146.631333850" Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.011471 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.012100 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:19.512073615 +0000 UTC m=+147.159758862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.025037 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" podStartSLOduration=121.025014099 podStartE2EDuration="2m1.025014099s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:19.023755012 +0000 UTC m=+146.671440079" watchObservedRunningTime="2025-10-02 09:39:19.025014099 +0000 UTC m=+146.672699166" Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.118712 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.118888 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:19.618862821 +0000 UTC m=+147.266547898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.119815 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.120403 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:19.620378785 +0000 UTC m=+147.268063852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.221383 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.221548 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:19.721506338 +0000 UTC m=+147.369191405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.221936 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.222510 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:19.722497806 +0000 UTC m=+147.370182873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.323445 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.323870 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:19.823854646 +0000 UTC m=+147.471539713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.428849 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.429708 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:19.929692695 +0000 UTC m=+147.577377762 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.530773 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.530983 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.030958872 +0000 UTC m=+147.678643939 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.531845 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.532286 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.03226509 +0000 UTC m=+147.679950157 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.636725 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.636910 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.136876013 +0000 UTC m=+147.784561080 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.637026 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.637466 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.13744979 +0000 UTC m=+147.785134867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.739001 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.739212 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.23918094 +0000 UTC m=+147.886866007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.740050 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.740415 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.240406746 +0000 UTC m=+147.888091813 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.840998 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.841236 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.341195349 +0000 UTC m=+147.988880436 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.841453 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.841914 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.341894329 +0000 UTC m=+147.989579586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.942393 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.942626 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.442588499 +0000 UTC m=+148.090273566 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.948749 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:19 crc kubenswrapper[4771]: E1002 09:39:19.949237 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.449212531 +0000 UTC m=+148.096897598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.983896 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" event={"ID":"e7769bc8-aad5-4e10-99f0-968f5b3e8b57","Type":"ContainerStarted","Data":"9e2d05894cac969f4d56f86e327441533d83e1d967ace312f1bccae713d0ea3f"} Oct 02 09:39:19 crc kubenswrapper[4771]: I1002 09:39:19.988561 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-n6z8c" event={"ID":"378a97ab-3ce3-4000-b15c-61c92f834207","Type":"ContainerStarted","Data":"48ba41c4811ce86eecb46224a152be80e170d9710efcf74c784291ad97ba6072"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.006071 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" event={"ID":"c65812c5-b7ed-434b-bdef-bbbe26fe5c69","Type":"ContainerStarted","Data":"20f0a6a0d7934c4538d30e7f3491203899646ff015600b1ae805be5b5682261f"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.006865 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7kpd" podStartSLOduration=122.006827276 podStartE2EDuration="2m2.006827276s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.004452938 +0000 UTC m=+147.652138005" watchObservedRunningTime="2025-10-02 09:39:20.006827276 +0000 UTC m=+147.654512343" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.036880 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" event={"ID":"d039518e-be81-4d51-9798-7043a3b653e0","Type":"ContainerStarted","Data":"283d19431c93e6dbfe852c2a338d8009e542b8302e2ab099a45aa2b02f07cbd7"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.041264 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6zk5s" podStartSLOduration=123.041234041 podStartE2EDuration="2m3.041234041s" podCreationTimestamp="2025-10-02 09:37:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.026574537 +0000 UTC m=+147.674259594" watchObservedRunningTime="2025-10-02 09:39:20.041234041 +0000 UTC m=+147.688919108" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.048600 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-449x6" event={"ID":"9b94af46-45c2-4f5c-9a16-53a1cedbe976","Type":"ContainerStarted","Data":"5196e939359380f277e99e6ccc055fee2fdd522981096c181e7ce4a9a12c811f"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.050215 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:20 crc kubenswrapper[4771]: E1002 09:39:20.050326 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.550298513 +0000 UTC m=+148.197983580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.050610 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:20 crc kubenswrapper[4771]: E1002 09:39:20.051157 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.551119916 +0000 UTC m=+148.198804983 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.059076 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7prrf" event={"ID":"a8ab108d-1536-402c-936a-b4689994b820","Type":"ContainerStarted","Data":"3bf0fa5fc3e49bb8e209846f84862a9b1b6f18eda88fee817a40213a2ced0c5d"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.059192 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7prrf" event={"ID":"a8ab108d-1536-402c-936a-b4689994b820","Type":"ContainerStarted","Data":"2cff547aab6fa1f14dc743b7945fcf7e219cd2eebfd99659d22c0b63e5885a92"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.074344 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-449x6" podStartSLOduration=6.074324887 podStartE2EDuration="6.074324887s" podCreationTimestamp="2025-10-02 09:39:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.073549545 +0000 UTC m=+147.721234612" watchObservedRunningTime="2025-10-02 09:39:20.074324887 +0000 UTC m=+147.722009954" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.094561 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nfr9d" event={"ID":"19247691-b8da-4344-91e1-57bc628fca0a","Type":"ContainerStarted","Data":"311134ce720985a0012dd2d1253cc92a4e97d5556f8ffb080e734a87682b362a"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.094653 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nfr9d" event={"ID":"19247691-b8da-4344-91e1-57bc628fca0a","Type":"ContainerStarted","Data":"80e5f57a74a8343e2999a29bcb28af35aeb21437bb5f519af8b09e51b7f40f0a"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.101996 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-z7f8f" event={"ID":"12fbfa54-08b0-41aa-ba11-80f84d2724cf","Type":"ContainerStarted","Data":"5d7922bfae2bff048f380779d0910a37f3129cafd8ef6f1030663f41c3dd4075"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.116204 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zpnmm" event={"ID":"e3fe7e38-10be-47d5-bd99-d01af7cbbf88","Type":"ContainerStarted","Data":"6c4e3ae0fb9cd5ab9f5b4122f05d9bcb6b7e744ab9e1b5a14bd5b58b6a2b08cb"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.134953 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" event={"ID":"961e712e-5760-4436-9be7-6bb6be2f8cf3","Type":"ContainerStarted","Data":"e5de3aea44f83ad887f2ecadd03eee62ced5a274592296cf8a4a617ba07e88e1"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.143741 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-nfr9d" podStartSLOduration=122.143706102 podStartE2EDuration="2m2.143706102s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.142992292 +0000 UTC m=+147.790677349" watchObservedRunningTime="2025-10-02 09:39:20.143706102 +0000 UTC m=+147.791391169" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.156213 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.157002 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" event={"ID":"46cb211b-6f4d-4484-8018-44c24c295287","Type":"ContainerStarted","Data":"4e5ffd7db58e8a6d947226bc4c7e0ec064e2810b97d3e322723d374a52628627"} Oct 02 09:39:20 crc kubenswrapper[4771]: E1002 09:39:20.163932 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.663906976 +0000 UTC m=+148.311592043 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.180226 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" event={"ID":"0524aac0-d101-4b59-b829-95390e99a159","Type":"ContainerStarted","Data":"33a8e8f25f4da29dd74bf68c9c03df4ac3c536b80d6e6db17f87667ae93399a6"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.180308 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" event={"ID":"0524aac0-d101-4b59-b829-95390e99a159","Type":"ContainerStarted","Data":"7451b1baaabfaf84c8887d0f66a420426cd7943adff2ac3fccea06961872ee66"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.192182 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-z7f8f" podStartSLOduration=122.192157743 podStartE2EDuration="2m2.192157743s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.191150524 +0000 UTC m=+147.838835591" watchObservedRunningTime="2025-10-02 09:39:20.192157743 +0000 UTC m=+147.839842810" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.202011 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" event={"ID":"c8cb8248-6e33-49aa-9422-f2e14c0e9015","Type":"ContainerStarted","Data":"6cf9643cfcc5c76d1b670f7215b3a182089b5809141a6d246bfed69e93e507be"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.213880 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" event={"ID":"88b88dc4-a78f-489f-b970-f70828c68f6d","Type":"ContainerStarted","Data":"b6fde8a1919db833d57aa1cf7ba6c3555bdab96eef7a15f37a79ce2b89ff2073"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.231947 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" event={"ID":"722c0898-dbc8-4f96-88f3-dcc328e61bd0","Type":"ContainerStarted","Data":"0d1a3461a22e2577bf9df9c559bc1e2c35096e300b556a6f89822c3b16dac723"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.261232 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:20 crc kubenswrapper[4771]: E1002 09:39:20.262493 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.762462295 +0000 UTC m=+148.410147362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.287546 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2" event={"ID":"e0904a9e-81f3-4849-9f7d-90e2485a218a","Type":"ContainerStarted","Data":"32f1bcc69949ca20a73ed0c751f2c5e6a754e62c556d85d8749ec2a0c7b6dd12"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.288736 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pz4l5" podStartSLOduration=122.288716054 podStartE2EDuration="2m2.288716054s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.287115667 +0000 UTC m=+147.934800734" watchObservedRunningTime="2025-10-02 09:39:20.288716054 +0000 UTC m=+147.936401121" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.290010 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-zpnmm" podStartSLOduration=7.290002951 podStartE2EDuration="7.290002951s" podCreationTimestamp="2025-10-02 09:39:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.222644084 +0000 UTC m=+147.870329151" watchObservedRunningTime="2025-10-02 09:39:20.290002951 +0000 UTC m=+147.937688018" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.296169 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" event={"ID":"220faef5-5c96-4a65-9692-183a34d12a44","Type":"ContainerStarted","Data":"1153eac35b388b2d064c7dad606f2a898cf8a30f22ebceb5c2c2c8cb5fd3a20e"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.296233 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" event={"ID":"220faef5-5c96-4a65-9692-183a34d12a44","Type":"ContainerStarted","Data":"aa7829c4aab2c688717f729c6563002a5c31280b1808dc194a2950a61db91d27"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.297202 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.299765 4771 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-p2qn6 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.299947 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" podUID="220faef5-5c96-4a65-9692-183a34d12a44" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.312682 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" event={"ID":"dd438f5d-c3ac-4a51-bc47-acd68fcc356a","Type":"ContainerStarted","Data":"8effe7c3b02f19f78c7bb8db02eb76189037b19b40b618504ad04aaea700cc6a"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.314025 4771 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-g92wn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.314103 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" podUID="dd438f5d-c3ac-4a51-bc47-acd68fcc356a" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.325009 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" event={"ID":"1cfc1793-db10-47a8-a152-0c86cc82be4c","Type":"ContainerStarted","Data":"3eec250664619fe65bf2e4606b866c258465a184306543fe3c8202e5f1c064ac"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.343533 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" event={"ID":"d632e53b-2327-40eb-9609-2449f74541c0","Type":"ContainerStarted","Data":"3ffb8715d1816a7ca71be2bcf15dc8f867b0831ae64d56c0f81e414d0ebdabec"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.349550 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" event={"ID":"b3217c27-f7c4-4b0f-9503-1968ce24eaba","Type":"ContainerStarted","Data":"fc9e5bf074d69dda1c8739c85526a434cfbe80ae1a2304462f09c3cd1f6b0cd7"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.362285 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm" event={"ID":"61b3178b-0376-4db5-8369-af20d1a30056","Type":"ContainerStarted","Data":"04c21d4e48c7c0e09f3de884420fe4479c067215d3e404c835a9a314f4116d98"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.363387 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:20 crc kubenswrapper[4771]: E1002 09:39:20.364332 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.864299318 +0000 UTC m=+148.511984535 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.364700 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:20 crc kubenswrapper[4771]: E1002 09:39:20.366534 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.866521492 +0000 UTC m=+148.514206749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.368217 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h8j9j" podStartSLOduration=122.36816994 podStartE2EDuration="2m2.36816994s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.336615258 +0000 UTC m=+147.984300325" watchObservedRunningTime="2025-10-02 09:39:20.36816994 +0000 UTC m=+148.015855007" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.369825 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6n2dp" podStartSLOduration=122.369814578 podStartE2EDuration="2m2.369814578s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.36088336 +0000 UTC m=+148.008568427" watchObservedRunningTime="2025-10-02 09:39:20.369814578 +0000 UTC m=+148.017499645" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.369849 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" event={"ID":"e3849bee-5c76-4308-8476-8abc1914d609","Type":"ContainerStarted","Data":"0cd4aee86a2ec1c56435e049edfb891fc6a1f4f04ad4d2bc82e4a965e8f0e312"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.395651 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-68c2v" podStartSLOduration=122.395633384 podStartE2EDuration="2m2.395633384s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.394580003 +0000 UTC m=+148.042265070" watchObservedRunningTime="2025-10-02 09:39:20.395633384 +0000 UTC m=+148.043318451" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.401721 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" event={"ID":"0ffcb835-7bd1-4cb3-b7cf-354595aa0b27","Type":"ContainerStarted","Data":"e295cee66b7c3b0d99ef4084777087789d29f4cdf85b686f780ab8a18e3f0a6d"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.401770 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" event={"ID":"0ffcb835-7bd1-4cb3-b7cf-354595aa0b27","Type":"ContainerStarted","Data":"65d6fea7a2b9e82d25ad889eec043b7e280701b1d4a1feb1df5ce261fec522e5"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.409569 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" event={"ID":"700cecb9-292a-4b4f-b244-528242ab782b","Type":"ContainerStarted","Data":"82baf9148dcbc94e0fa2e90ae7fd80104ce7dbcf5cca9dbfaa8e9e83bc835857"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.409618 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" event={"ID":"700cecb9-292a-4b4f-b244-528242ab782b","Type":"ContainerStarted","Data":"1584ea148af6fd36f3c3bb7fc16f30430db50a46534abff3e7b877c29c2c5a41"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.412633 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" event={"ID":"4917c5c7-bed1-46fc-ab93-7f9959cc0b48","Type":"ContainerStarted","Data":"45327827580e7aea074eef6c56d6b8e95664cb9cc7e80802eb329ec37ee45978"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.414597 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jb4nn" event={"ID":"9fd48e0d-81e5-49ed-8ced-4c382f96f831","Type":"ContainerStarted","Data":"2a66d08edc2253b2acc35d3533aefa8d032b46b338de63ad15d21da807ad71be"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.431996 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rrb82" podStartSLOduration=122.431978504 podStartE2EDuration="2m2.431978504s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.431253853 +0000 UTC m=+148.078938920" watchObservedRunningTime="2025-10-02 09:39:20.431978504 +0000 UTC m=+148.079663561" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.438358 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" event={"ID":"671d32ff-3e87-49bd-a4ec-60a495671813","Type":"ContainerStarted","Data":"de065b48a0324860435a19038c86e31520939cf1a286fc8247e3e5d225e493ab"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.439022 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.444473 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" event={"ID":"fefa105a-53b5-45a0-ac11-f3078d78ae59","Type":"ContainerStarted","Data":"24ac4dbd155e9435ab6b49ccbe17b2875b9cd305455f16316b8f2bf0a1ea32f8"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.454089 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" event={"ID":"e09f2d35-5bde-444c-866d-1376247d4659","Type":"ContainerStarted","Data":"633861a9f611c3118c907b03f0379cd2d5c5df9aa2226bd6fe9f56bb57798613"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.456070 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2" podStartSLOduration=122.45604642 podStartE2EDuration="2m2.45604642s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.455118883 +0000 UTC m=+148.102803950" watchObservedRunningTime="2025-10-02 09:39:20.45604642 +0000 UTC m=+148.103731487" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.461469 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" event={"ID":"5c65c130-1f41-42f8-be1a-9c8f17bf16dc","Type":"ContainerStarted","Data":"ea79706ce5283488be0840742616fa2605002ad5a3acb7809e6bddbe76075725"} Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.464644 4771 patch_prober.go:28] interesting pod/console-operator-58897d9998-8dj62 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.464705 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-8dj62" podUID="d1366a10-f981-4c9a-aecd-415273b5b714" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.464736 4771 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-p65v7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.464806 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" podUID="b8f91205-c6e3-4f73-ac1b-21a7e79620c1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.465327 4771 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-4wzs8 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.465398 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" podUID="a833cc23-a6bc-4d07-95bf-d745791879d6" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.465831 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.465891 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.466023 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:20 crc kubenswrapper[4771]: E1002 09:39:20.472735 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.972680051 +0000 UTC m=+148.620365118 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.473287 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:20 crc kubenswrapper[4771]: E1002 09:39:20.481814 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:20.981774254 +0000 UTC m=+148.629459331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.504437 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j67pl" podStartSLOduration=122.504410288 podStartE2EDuration="2m2.504410288s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.499507496 +0000 UTC m=+148.147192563" watchObservedRunningTime="2025-10-02 09:39:20.504410288 +0000 UTC m=+148.152095375" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.572353 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xbvnm" podStartSLOduration=122.572317521 podStartE2EDuration="2m2.572317521s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.572044513 +0000 UTC m=+148.219729580" watchObservedRunningTime="2025-10-02 09:39:20.572317521 +0000 UTC m=+148.220002588" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.574145 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" podStartSLOduration=122.574117743 podStartE2EDuration="2m2.574117743s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.539692308 +0000 UTC m=+148.187377375" watchObservedRunningTime="2025-10-02 09:39:20.574117743 +0000 UTC m=+148.221802820" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.574908 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:20 crc kubenswrapper[4771]: E1002 09:39:20.576982 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:21.076965035 +0000 UTC m=+148.724650102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.607803 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-l4x98" podStartSLOduration=122.607785966 podStartE2EDuration="2m2.607785966s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.606160589 +0000 UTC m=+148.253845676" watchObservedRunningTime="2025-10-02 09:39:20.607785966 +0000 UTC m=+148.255471033" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.677724 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:20 crc kubenswrapper[4771]: E1002 09:39:20.678269 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:21.178252322 +0000 UTC m=+148.825937389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.683776 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" podStartSLOduration=123.683752451 podStartE2EDuration="2m3.683752451s" podCreationTimestamp="2025-10-02 09:37:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.628322989 +0000 UTC m=+148.276008056" watchObservedRunningTime="2025-10-02 09:39:20.683752451 +0000 UTC m=+148.331437518" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.684908 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" podStartSLOduration=122.684900515 podStartE2EDuration="2m2.684900515s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.683563926 +0000 UTC m=+148.331249003" watchObservedRunningTime="2025-10-02 09:39:20.684900515 +0000 UTC m=+148.332585582" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.719222 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" podStartSLOduration=122.719204526 podStartE2EDuration="2m2.719204526s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.717834067 +0000 UTC m=+148.365519134" watchObservedRunningTime="2025-10-02 09:39:20.719204526 +0000 UTC m=+148.366889593" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.742041 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" podStartSLOduration=122.742024856 podStartE2EDuration="2m2.742024856s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.739213205 +0000 UTC m=+148.386898272" watchObservedRunningTime="2025-10-02 09:39:20.742024856 +0000 UTC m=+148.389709923" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.775667 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r8c2x" podStartSLOduration=122.775643517 podStartE2EDuration="2m2.775643517s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.773358901 +0000 UTC m=+148.421043978" watchObservedRunningTime="2025-10-02 09:39:20.775643517 +0000 UTC m=+148.423328584" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.778461 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:20 crc kubenswrapper[4771]: E1002 09:39:20.778963 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:21.278942483 +0000 UTC m=+148.926627550 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.800713 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-vz4s6" podStartSLOduration=122.800691061 podStartE2EDuration="2m2.800691061s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:20.800222948 +0000 UTC m=+148.447908015" watchObservedRunningTime="2025-10-02 09:39:20.800691061 +0000 UTC m=+148.448376128" Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.880470 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:20 crc kubenswrapper[4771]: E1002 09:39:20.880889 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:21.380873869 +0000 UTC m=+149.028558946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:20 crc kubenswrapper[4771]: I1002 09:39:20.981290 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:20 crc kubenswrapper[4771]: E1002 09:39:20.981698 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:21.481679743 +0000 UTC m=+149.129364810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.057938 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.057997 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.058013 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.059678 4771 patch_prober.go:28] interesting pod/router-default-5444994796-z7f8f container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.059732 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z7f8f" podUID="12fbfa54-08b0-41aa-ba11-80f84d2724cf" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.060011 4771 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-rqp4t container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.060031 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" podUID="fefa105a-53b5-45a0-ac11-f3078d78ae59" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.083077 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:21 crc kubenswrapper[4771]: E1002 09:39:21.083566 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:21.583549796 +0000 UTC m=+149.231234863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.183892 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:21 crc kubenswrapper[4771]: E1002 09:39:21.184334 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:21.684318638 +0000 UTC m=+149.332003705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.285515 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:21 crc kubenswrapper[4771]: E1002 09:39:21.285823 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:21.785809082 +0000 UTC m=+149.433494149 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.290263 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.386454 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:21 crc kubenswrapper[4771]: E1002 09:39:21.386623 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:21.886589395 +0000 UTC m=+149.534274462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.386936 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:21 crc kubenswrapper[4771]: E1002 09:39:21.387521 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:21.887508871 +0000 UTC m=+149.535193938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.475857 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" event={"ID":"722c0898-dbc8-4f96-88f3-dcc328e61bd0","Type":"ContainerStarted","Data":"4aeca16cf7d2900bbbf4dc69a91925f41e14d07ce39ce126394c71616b5d0660"} Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.481609 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" event={"ID":"d039518e-be81-4d51-9798-7043a3b653e0","Type":"ContainerStarted","Data":"d0b6d6bbef91d7c1933ec777343a433d4d0b5642fb52539e5018d86101854b23"} Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.485579 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" event={"ID":"961e712e-5760-4436-9be7-6bb6be2f8cf3","Type":"ContainerStarted","Data":"5c54c945058d16f591de9d28b8d704c982b837f1cb87a07e67e937d552e5db7d"} Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.487512 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rlhp2" event={"ID":"e0904a9e-81f3-4849-9f7d-90e2485a218a","Type":"ContainerStarted","Data":"340f6f2db7ac7a5997c859faad7bba652760582663d2efe089a0db153436ea92"} Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.487818 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:21 crc kubenswrapper[4771]: E1002 09:39:21.488024 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:21.987991526 +0000 UTC m=+149.635676593 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.488427 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:21 crc kubenswrapper[4771]: E1002 09:39:21.488687 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:21.988666325 +0000 UTC m=+149.636351402 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.491258 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7prrf" event={"ID":"a8ab108d-1536-402c-936a-b4689994b820","Type":"ContainerStarted","Data":"ca5e9c1cb5b0fd534d35be915fd50e6cc61d369af1506ebb5f031e52edb5b0dd"} Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.494489 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" event={"ID":"0524aac0-d101-4b59-b829-95390e99a159","Type":"ContainerStarted","Data":"c95cb73767b8726388fc11fd090ed1f929fa6a9d51cef78e58fa0c48e95777b3"} Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.494997 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.497702 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-n6z8c" event={"ID":"378a97ab-3ce3-4000-b15c-61c92f834207","Type":"ContainerStarted","Data":"b0ba4c26e79b883da01a9695abf2de7b73f6c1fe7b17d8a1ba7c463559fac7a5"} Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.498200 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-n6z8c" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.511520 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jb4nn" event={"ID":"9fd48e0d-81e5-49ed-8ced-4c382f96f831","Type":"ContainerStarted","Data":"c04fb03e86b0245d0b145242d498db9ab7d371cacb82e634a4621ffb08bffe94"} Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.515117 4771 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-p2qn6 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.515232 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" podUID="220faef5-5c96-4a65-9692-183a34d12a44" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.516767 4771 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cx2j5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.516808 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" podUID="5c65c130-1f41-42f8-be1a-9c8f17bf16dc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.518070 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnvsk" podStartSLOduration=123.518052344 podStartE2EDuration="2m3.518052344s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:21.505957295 +0000 UTC m=+149.153642362" watchObservedRunningTime="2025-10-02 09:39:21.518052344 +0000 UTC m=+149.165737411" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.519793 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.529203 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.571797 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wzs8" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.590718 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:21 crc kubenswrapper[4771]: E1002 09:39:21.591146 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.091088895 +0000 UTC m=+149.738774002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.626648 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-7prrf" podStartSLOduration=123.626626793 podStartE2EDuration="2m3.626626793s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:21.572299832 +0000 UTC m=+149.219984899" watchObservedRunningTime="2025-10-02 09:39:21.626626793 +0000 UTC m=+149.274311860" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.628212 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" podStartSLOduration=124.628206008 podStartE2EDuration="2m4.628206008s" podCreationTimestamp="2025-10-02 09:37:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:21.625764898 +0000 UTC m=+149.273449975" watchObservedRunningTime="2025-10-02 09:39:21.628206008 +0000 UTC m=+149.275891075" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.685177 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" podStartSLOduration=123.685158694 podStartE2EDuration="2m3.685158694s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:21.684543607 +0000 UTC m=+149.332228674" watchObservedRunningTime="2025-10-02 09:39:21.685158694 +0000 UTC m=+149.332843761" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.685695 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-n6z8c" podStartSLOduration=8.68568753 podStartE2EDuration="8.68568753s" podCreationTimestamp="2025-10-02 09:39:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:21.664795976 +0000 UTC m=+149.312481043" watchObservedRunningTime="2025-10-02 09:39:21.68568753 +0000 UTC m=+149.333372607" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.693759 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:21 crc kubenswrapper[4771]: E1002 09:39:21.694161 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.194113853 +0000 UTC m=+149.841798920 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.764896 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-jb4nn" podStartSLOduration=123.764877528 podStartE2EDuration="2m3.764877528s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:21.763046986 +0000 UTC m=+149.410732053" watchObservedRunningTime="2025-10-02 09:39:21.764877528 +0000 UTC m=+149.412562595" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.765977 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fw2" podStartSLOduration=123.76597174 podStartE2EDuration="2m3.76597174s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:21.721518025 +0000 UTC m=+149.369203112" watchObservedRunningTime="2025-10-02 09:39:21.76597174 +0000 UTC m=+149.413656807" Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.794537 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:21 crc kubenswrapper[4771]: E1002 09:39:21.794925 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.294908956 +0000 UTC m=+149.942594023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.896800 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:21 crc kubenswrapper[4771]: E1002 09:39:21.897157 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.397138861 +0000 UTC m=+150.044823938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.998172 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:21 crc kubenswrapper[4771]: E1002 09:39:21.998268 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.498249734 +0000 UTC m=+150.145934801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:21 crc kubenswrapper[4771]: I1002 09:39:21.998344 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:21 crc kubenswrapper[4771]: E1002 09:39:21.998657 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.498648425 +0000 UTC m=+150.146333492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.060796 4771 patch_prober.go:28] interesting pod/router-default-5444994796-z7f8f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:39:22 crc kubenswrapper[4771]: [-]has-synced failed: reason withheld Oct 02 09:39:22 crc kubenswrapper[4771]: [+]process-running ok Oct 02 09:39:22 crc kubenswrapper[4771]: healthz check failed Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.060862 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z7f8f" podUID="12fbfa54-08b0-41aa-ba11-80f84d2724cf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.099019 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.099210 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.599184451 +0000 UTC m=+150.246869518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.099334 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.099674 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.599664445 +0000 UTC m=+150.247349512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.200579 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.200779 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.700750977 +0000 UTC m=+150.348436044 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.200840 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.201199 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.701190189 +0000 UTC m=+150.348875256 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.302162 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.302387 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.802359163 +0000 UTC m=+150.450044230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.302536 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.302861 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.802847838 +0000 UTC m=+150.450532895 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.403524 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.403824 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.903791585 +0000 UTC m=+150.551476662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.403953 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.404351 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:22.904343541 +0000 UTC m=+150.552028608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.449209 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.505740 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.505930 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.005903997 +0000 UTC m=+150.653589064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.506005 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.506094 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.506116 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.506184 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.506554 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.006540705 +0000 UTC m=+150.654225772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.506957 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.511452 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.512201 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.529532 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" event={"ID":"e3849bee-5c76-4308-8476-8abc1914d609","Type":"ContainerStarted","Data":"ecf2fad0d82d6edbe05f9fc69c51785dc9c6f803d090b8dd8cd0dd983dd049bf"} Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.532286 4771 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cx2j5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.532331 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" podUID="5c65c130-1f41-42f8-be1a-9c8f17bf16dc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.548709 4771 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-zlrjz container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.548766 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" podUID="671d32ff-3e87-49bd-a4ec-60a495671813" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.553377 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p2qn6" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.607272 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.607923 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.609530 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.109501181 +0000 UTC m=+150.757186408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.651951 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.712485 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.719763 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.720070 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.720385 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.220372175 +0000 UTC m=+150.868057242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.730343 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.821804 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.822084 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.322053634 +0000 UTC m=+150.969738701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:22 crc kubenswrapper[4771]: I1002 09:39:22.923471 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:22 crc kubenswrapper[4771]: E1002 09:39:22.923911 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.423894638 +0000 UTC m=+151.071579705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.025303 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:23 crc kubenswrapper[4771]: E1002 09:39:23.025538 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.525490924 +0000 UTC m=+151.173175991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.025609 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:23 crc kubenswrapper[4771]: E1002 09:39:23.025953 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.525933857 +0000 UTC m=+151.173618924 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.060394 4771 patch_prober.go:28] interesting pod/router-default-5444994796-z7f8f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:39:23 crc kubenswrapper[4771]: [-]has-synced failed: reason withheld Oct 02 09:39:23 crc kubenswrapper[4771]: [+]process-running ok Oct 02 09:39:23 crc kubenswrapper[4771]: healthz check failed Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.060733 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z7f8f" podUID="12fbfa54-08b0-41aa-ba11-80f84d2724cf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.127407 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:23 crc kubenswrapper[4771]: E1002 09:39:23.127735 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.627720859 +0000 UTC m=+151.275405926 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.229000 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:23 crc kubenswrapper[4771]: E1002 09:39:23.229433 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.729417348 +0000 UTC m=+151.377102425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.330678 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:23 crc kubenswrapper[4771]: E1002 09:39:23.330927 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.83085661 +0000 UTC m=+151.478541677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.330977 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:23 crc kubenswrapper[4771]: E1002 09:39:23.331486 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.831470218 +0000 UTC m=+151.479155465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.434983 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:23 crc kubenswrapper[4771]: E1002 09:39:23.435254 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.935230397 +0000 UTC m=+151.582915464 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.435373 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:23 crc kubenswrapper[4771]: E1002 09:39:23.435750 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:23.935739122 +0000 UTC m=+151.583424189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.536382 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:23 crc kubenswrapper[4771]: E1002 09:39:23.536757 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:24.036740931 +0000 UTC m=+151.684425988 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.544895 4771 generic.go:334] "Generic (PLEG): container finished" podID="e09f2d35-5bde-444c-866d-1376247d4659" containerID="633861a9f611c3118c907b03f0379cd2d5c5df9aa2226bd6fe9f56bb57798613" exitCode=0 Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.545787 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" event={"ID":"e09f2d35-5bde-444c-866d-1376247d4659","Type":"ContainerDied","Data":"633861a9f611c3118c907b03f0379cd2d5c5df9aa2226bd6fe9f56bb57798613"} Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.640002 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:23 crc kubenswrapper[4771]: E1002 09:39:23.641959 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:24.141932371 +0000 UTC m=+151.789617438 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.746928 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:23 crc kubenswrapper[4771]: E1002 09:39:23.747358 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:24.247335738 +0000 UTC m=+151.895020805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.849173 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:23 crc kubenswrapper[4771]: E1002 09:39:23.849601 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:24.349586263 +0000 UTC m=+151.997271330 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:23 crc kubenswrapper[4771]: I1002 09:39:23.953704 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:23 crc kubenswrapper[4771]: E1002 09:39:23.954108 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:24.454090724 +0000 UTC m=+152.101775791 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.055614 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:24 crc kubenswrapper[4771]: E1002 09:39:24.056641 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:24.556623287 +0000 UTC m=+152.204308354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.061831 4771 patch_prober.go:28] interesting pod/router-default-5444994796-z7f8f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:39:24 crc kubenswrapper[4771]: [-]has-synced failed: reason withheld Oct 02 09:39:24 crc kubenswrapper[4771]: [+]process-running ok Oct 02 09:39:24 crc kubenswrapper[4771]: healthz check failed Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.061895 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z7f8f" podUID="12fbfa54-08b0-41aa-ba11-80f84d2724cf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.121770 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l5tmt"] Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.127635 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.133716 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.153793 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l5tmt"] Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.159747 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:24 crc kubenswrapper[4771]: E1002 09:39:24.160065 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:24.660048257 +0000 UTC m=+152.307733324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.264086 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fcbdb41-18a3-4977-aa8b-28f5af63576f-catalog-content\") pod \"certified-operators-l5tmt\" (UID: \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\") " pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.264394 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.264486 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fcbdb41-18a3-4977-aa8b-28f5af63576f-utilities\") pod \"certified-operators-l5tmt\" (UID: \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\") " pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.264751 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bpzm\" (UniqueName: \"kubernetes.io/projected/8fcbdb41-18a3-4977-aa8b-28f5af63576f-kube-api-access-8bpzm\") pod \"certified-operators-l5tmt\" (UID: \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\") " pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:39:24 crc kubenswrapper[4771]: E1002 09:39:24.264885 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:24.764841485 +0000 UTC m=+152.412526552 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.302235 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lmpwh"] Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.304054 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.312395 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.347387 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lmpwh"] Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.365873 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.366263 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bpzm\" (UniqueName: \"kubernetes.io/projected/8fcbdb41-18a3-4977-aa8b-28f5af63576f-kube-api-access-8bpzm\") pod \"certified-operators-l5tmt\" (UID: \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\") " pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.366298 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fcbdb41-18a3-4977-aa8b-28f5af63576f-catalog-content\") pod \"certified-operators-l5tmt\" (UID: \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\") " pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.366324 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d198e457-892f-40a9-9c39-46e4e67b79a8-catalog-content\") pod \"community-operators-lmpwh\" (UID: \"d198e457-892f-40a9-9c39-46e4e67b79a8\") " pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.366346 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm85t\" (UniqueName: \"kubernetes.io/projected/d198e457-892f-40a9-9c39-46e4e67b79a8-kube-api-access-bm85t\") pod \"community-operators-lmpwh\" (UID: \"d198e457-892f-40a9-9c39-46e4e67b79a8\") " pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.366363 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d198e457-892f-40a9-9c39-46e4e67b79a8-utilities\") pod \"community-operators-lmpwh\" (UID: \"d198e457-892f-40a9-9c39-46e4e67b79a8\") " pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.366422 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fcbdb41-18a3-4977-aa8b-28f5af63576f-utilities\") pod \"certified-operators-l5tmt\" (UID: \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\") " pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.366815 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fcbdb41-18a3-4977-aa8b-28f5af63576f-utilities\") pod \"certified-operators-l5tmt\" (UID: \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\") " pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:39:24 crc kubenswrapper[4771]: E1002 09:39:24.367028 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:24.867006398 +0000 UTC m=+152.514691465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.367676 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fcbdb41-18a3-4977-aa8b-28f5af63576f-catalog-content\") pod \"certified-operators-l5tmt\" (UID: \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\") " pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.416545 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bpzm\" (UniqueName: \"kubernetes.io/projected/8fcbdb41-18a3-4977-aa8b-28f5af63576f-kube-api-access-8bpzm\") pod \"certified-operators-l5tmt\" (UID: \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\") " pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.474170 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d198e457-892f-40a9-9c39-46e4e67b79a8-catalog-content\") pod \"community-operators-lmpwh\" (UID: \"d198e457-892f-40a9-9c39-46e4e67b79a8\") " pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.474227 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm85t\" (UniqueName: \"kubernetes.io/projected/d198e457-892f-40a9-9c39-46e4e67b79a8-kube-api-access-bm85t\") pod \"community-operators-lmpwh\" (UID: \"d198e457-892f-40a9-9c39-46e4e67b79a8\") " pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.474252 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d198e457-892f-40a9-9c39-46e4e67b79a8-utilities\") pod \"community-operators-lmpwh\" (UID: \"d198e457-892f-40a9-9c39-46e4e67b79a8\") " pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.474292 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:24 crc kubenswrapper[4771]: E1002 09:39:24.474675 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:24.9746602 +0000 UTC m=+152.622345267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.475211 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d198e457-892f-40a9-9c39-46e4e67b79a8-catalog-content\") pod \"community-operators-lmpwh\" (UID: \"d198e457-892f-40a9-9c39-46e4e67b79a8\") " pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.475302 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d198e457-892f-40a9-9c39-46e4e67b79a8-utilities\") pod \"community-operators-lmpwh\" (UID: \"d198e457-892f-40a9-9c39-46e4e67b79a8\") " pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.528407 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zxgdb"] Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.529421 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.536012 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm85t\" (UniqueName: \"kubernetes.io/projected/d198e457-892f-40a9-9c39-46e4e67b79a8-kube-api-access-bm85t\") pod \"community-operators-lmpwh\" (UID: \"d198e457-892f-40a9-9c39-46e4e67b79a8\") " pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.536549 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.544867 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zxgdb"] Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.577687 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.577913 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5107994-4460-4ac0-b975-6ba0fc73a542-catalog-content\") pod \"certified-operators-zxgdb\" (UID: \"d5107994-4460-4ac0-b975-6ba0fc73a542\") " pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.577952 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5tgt\" (UniqueName: \"kubernetes.io/projected/d5107994-4460-4ac0-b975-6ba0fc73a542-kube-api-access-b5tgt\") pod \"certified-operators-zxgdb\" (UID: \"d5107994-4460-4ac0-b975-6ba0fc73a542\") " pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.577969 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5107994-4460-4ac0-b975-6ba0fc73a542-utilities\") pod \"certified-operators-zxgdb\" (UID: \"d5107994-4460-4ac0-b975-6ba0fc73a542\") " pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:39:24 crc kubenswrapper[4771]: E1002 09:39:24.578081 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:25.078065719 +0000 UTC m=+152.725750776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.604567 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"52b91972418f6e955aee5301ac13a0e0bbe84ffbdbcc81ad3e74ab50d094b4d1"} Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.604617 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2bc1b7a8b63f90c1257191249c8aa926358e5fe2b80325445497371f99abcb82"} Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.613972 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"90f8b626eedb65f305f861b35f43d98de794ecc0e48063583e0167f65ca4691c"} Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.631460 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f74009b20b5c874063dc4887d5fce8cb1b43809607c6d4554b833fdcc5bf7a76"} Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.631509 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b1cb6fdcc3035cfca0cb351ca5c82cc871f7d545f8f02c1f619a6467df4ad713"} Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.632289 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.636022 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.678680 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5107994-4460-4ac0-b975-6ba0fc73a542-utilities\") pod \"certified-operators-zxgdb\" (UID: \"d5107994-4460-4ac0-b975-6ba0fc73a542\") " pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.678720 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5tgt\" (UniqueName: \"kubernetes.io/projected/d5107994-4460-4ac0-b975-6ba0fc73a542-kube-api-access-b5tgt\") pod \"certified-operators-zxgdb\" (UID: \"d5107994-4460-4ac0-b975-6ba0fc73a542\") " pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.678794 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.678825 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5107994-4460-4ac0-b975-6ba0fc73a542-catalog-content\") pod \"certified-operators-zxgdb\" (UID: \"d5107994-4460-4ac0-b975-6ba0fc73a542\") " pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.680329 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5107994-4460-4ac0-b975-6ba0fc73a542-catalog-content\") pod \"certified-operators-zxgdb\" (UID: \"d5107994-4460-4ac0-b975-6ba0fc73a542\") " pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:39:24 crc kubenswrapper[4771]: E1002 09:39:24.681275 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:25.18126183 +0000 UTC m=+152.828946897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.683262 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5107994-4460-4ac0-b975-6ba0fc73a542-utilities\") pod \"certified-operators-zxgdb\" (UID: \"d5107994-4460-4ac0-b975-6ba0fc73a542\") " pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.704111 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k2gjk"] Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.705592 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.727516 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k2gjk"] Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.744301 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5tgt\" (UniqueName: \"kubernetes.io/projected/d5107994-4460-4ac0-b975-6ba0fc73a542-kube-api-access-b5tgt\") pod \"certified-operators-zxgdb\" (UID: \"d5107994-4460-4ac0-b975-6ba0fc73a542\") " pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.790093 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:24 crc kubenswrapper[4771]: E1002 09:39:24.790284 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:25.29025095 +0000 UTC m=+152.937936037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.790324 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d01d796c-0c02-4c8d-b423-e32fc4b187d6-catalog-content\") pod \"community-operators-k2gjk\" (UID: \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\") " pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.790366 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m47qg\" (UniqueName: \"kubernetes.io/projected/d01d796c-0c02-4c8d-b423-e32fc4b187d6-kube-api-access-m47qg\") pod \"community-operators-k2gjk\" (UID: \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\") " pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.790471 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.790505 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d01d796c-0c02-4c8d-b423-e32fc4b187d6-utilities\") pod \"community-operators-k2gjk\" (UID: \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\") " pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:39:24 crc kubenswrapper[4771]: E1002 09:39:24.790836 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:25.290821247 +0000 UTC m=+152.938506314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.856104 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.891799 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:24 crc kubenswrapper[4771]: E1002 09:39:24.892420 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:25.392391703 +0000 UTC m=+153.040076770 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.892484 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.892570 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d01d796c-0c02-4c8d-b423-e32fc4b187d6-utilities\") pod \"community-operators-k2gjk\" (UID: \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\") " pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.892600 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d01d796c-0c02-4c8d-b423-e32fc4b187d6-catalog-content\") pod \"community-operators-k2gjk\" (UID: \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\") " pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.892661 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m47qg\" (UniqueName: \"kubernetes.io/projected/d01d796c-0c02-4c8d-b423-e32fc4b187d6-kube-api-access-m47qg\") pod \"community-operators-k2gjk\" (UID: \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\") " pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:39:24 crc kubenswrapper[4771]: E1002 09:39:24.893802 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:25.393788793 +0000 UTC m=+153.041473860 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.894322 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d01d796c-0c02-4c8d-b423-e32fc4b187d6-utilities\") pod \"community-operators-k2gjk\" (UID: \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\") " pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.897473 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d01d796c-0c02-4c8d-b423-e32fc4b187d6-catalog-content\") pod \"community-operators-k2gjk\" (UID: \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\") " pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.924730 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zlrjz" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.939033 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m47qg\" (UniqueName: \"kubernetes.io/projected/d01d796c-0c02-4c8d-b423-e32fc4b187d6-kube-api-access-m47qg\") pod \"community-operators-k2gjk\" (UID: \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\") " pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.993894 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:24 crc kubenswrapper[4771]: E1002 09:39:24.994203 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:25.494185695 +0000 UTC m=+153.141870762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:24 crc kubenswrapper[4771]: I1002 09:39:24.994584 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:24 crc kubenswrapper[4771]: E1002 09:39:24.995051 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:25.495031899 +0000 UTC m=+153.142716966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.043659 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.097460 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:25 crc kubenswrapper[4771]: E1002 09:39:25.097860 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:25.597842211 +0000 UTC m=+153.245527278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.113406 4771 patch_prober.go:28] interesting pod/router-default-5444994796-z7f8f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:39:25 crc kubenswrapper[4771]: [-]has-synced failed: reason withheld Oct 02 09:39:25 crc kubenswrapper[4771]: [+]process-running ok Oct 02 09:39:25 crc kubenswrapper[4771]: healthz check failed Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.121622 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z7f8f" podUID="12fbfa54-08b0-41aa-ba11-80f84d2724cf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.203500 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:25 crc kubenswrapper[4771]: E1002 09:39:25.203836 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:25.703823924 +0000 UTC m=+153.351508991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.265324 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l5tmt"] Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.304017 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:25 crc kubenswrapper[4771]: E1002 09:39:25.304538 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:25.804517614 +0000 UTC m=+153.452202681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.322867 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.361642 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 09:39:25 crc kubenswrapper[4771]: E1002 09:39:25.361917 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e09f2d35-5bde-444c-866d-1376247d4659" containerName="collect-profiles" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.361930 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e09f2d35-5bde-444c-866d-1376247d4659" containerName="collect-profiles" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.362031 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="e09f2d35-5bde-444c-866d-1376247d4659" containerName="collect-profiles" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.371334 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.394393 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.399687 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.404756 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e09f2d35-5bde-444c-866d-1376247d4659-secret-volume\") pod \"e09f2d35-5bde-444c-866d-1376247d4659\" (UID: \"e09f2d35-5bde-444c-866d-1376247d4659\") " Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.404865 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xqgk\" (UniqueName: \"kubernetes.io/projected/e09f2d35-5bde-444c-866d-1376247d4659-kube-api-access-6xqgk\") pod \"e09f2d35-5bde-444c-866d-1376247d4659\" (UID: \"e09f2d35-5bde-444c-866d-1376247d4659\") " Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.405048 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e09f2d35-5bde-444c-866d-1376247d4659-config-volume\") pod \"e09f2d35-5bde-444c-866d-1376247d4659\" (UID: \"e09f2d35-5bde-444c-866d-1376247d4659\") " Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.405267 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:25 crc kubenswrapper[4771]: E1002 09:39:25.405722 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:25.905705229 +0000 UTC m=+153.553390296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.417160 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.419732 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e09f2d35-5bde-444c-866d-1376247d4659-config-volume" (OuterVolumeSpecName: "config-volume") pod "e09f2d35-5bde-444c-866d-1376247d4659" (UID: "e09f2d35-5bde-444c-866d-1376247d4659"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.427596 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lmpwh"] Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.428006 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e09f2d35-5bde-444c-866d-1376247d4659-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e09f2d35-5bde-444c-866d-1376247d4659" (UID: "e09f2d35-5bde-444c-866d-1376247d4659"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.429537 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e09f2d35-5bde-444c-866d-1376247d4659-kube-api-access-6xqgk" (OuterVolumeSpecName: "kube-api-access-6xqgk") pod "e09f2d35-5bde-444c-866d-1376247d4659" (UID: "e09f2d35-5bde-444c-866d-1376247d4659"). InnerVolumeSpecName "kube-api-access-6xqgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.509610 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:25 crc kubenswrapper[4771]: E1002 09:39:25.517169 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:26.017112569 +0000 UTC m=+153.664797646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.517434 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a072c8af-e1fb-4af2-98ba-e023152b4cfd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a072c8af-e1fb-4af2-98ba-e023152b4cfd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.517527 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a072c8af-e1fb-4af2-98ba-e023152b4cfd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a072c8af-e1fb-4af2-98ba-e023152b4cfd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.517630 4771 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e09f2d35-5bde-444c-866d-1376247d4659-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.520565 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xqgk\" (UniqueName: \"kubernetes.io/projected/e09f2d35-5bde-444c-866d-1376247d4659-kube-api-access-6xqgk\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.520683 4771 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e09f2d35-5bde-444c-866d-1376247d4659-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.550747 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k2gjk"] Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.597589 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zxgdb"] Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.623417 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a072c8af-e1fb-4af2-98ba-e023152b4cfd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a072c8af-e1fb-4af2-98ba-e023152b4cfd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.623531 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.623635 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a072c8af-e1fb-4af2-98ba-e023152b4cfd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a072c8af-e1fb-4af2-98ba-e023152b4cfd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.623757 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a072c8af-e1fb-4af2-98ba-e023152b4cfd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a072c8af-e1fb-4af2-98ba-e023152b4cfd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:39:25 crc kubenswrapper[4771]: E1002 09:39:25.624296 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:26.124273726 +0000 UTC m=+153.771958943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.646830 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2gjk" event={"ID":"d01d796c-0c02-4c8d-b423-e32fc4b187d6","Type":"ContainerStarted","Data":"15d56732786d0ff7d14712b3edce8421a246eddd28446e24712b343a510b79e0"} Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.658104 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"501afd39cbf7379a7e274982b76b427ba34d1e064f4f999e6377304921e33457"} Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.659088 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a072c8af-e1fb-4af2-98ba-e023152b4cfd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a072c8af-e1fb-4af2-98ba-e023152b4cfd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.666350 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" event={"ID":"e3849bee-5c76-4308-8476-8abc1914d609","Type":"ContainerStarted","Data":"751fae43ed16a2face87d7a65c46e31da0e8d011ec2a5b0cb3b6f948931d75c9"} Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.669170 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lmpwh" event={"ID":"d198e457-892f-40a9-9c39-46e4e67b79a8","Type":"ContainerStarted","Data":"140977a6f77b0f00975bca943d1141919673b54b23bf5828b010fde8b71b2655"} Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.682509 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.682539 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r" event={"ID":"e09f2d35-5bde-444c-866d-1376247d4659","Type":"ContainerDied","Data":"a6cba2732c1f6e1e34d54df58abcebab9fb016f36dab507d6da83f6f9612a29c"} Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.682673 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6cba2732c1f6e1e34d54df58abcebab9fb016f36dab507d6da83f6f9612a29c" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.710300 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.726958 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:25 crc kubenswrapper[4771]: E1002 09:39:25.727841 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:26.227806999 +0000 UTC m=+153.875492066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.742671 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l5tmt" event={"ID":"8fcbdb41-18a3-4977-aa8b-28f5af63576f","Type":"ContainerStarted","Data":"56ca009950917e386f52ba1ea39d5f765fd9a67915a66cb072caed9c9024f99b"} Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.829842 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:25 crc kubenswrapper[4771]: E1002 09:39:25.830218 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:26.330204308 +0000 UTC m=+153.977889375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.883025 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.884834 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.896399 4771 patch_prober.go:28] interesting pod/console-f9d7485db-r664s container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.896462 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-r664s" podUID="67d2ef8d-5a0c-48fa-8233-9bb23e794da0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.917341 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xtdpd" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.930875 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:25 crc kubenswrapper[4771]: E1002 09:39:25.931056 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:26.431027213 +0000 UTC m=+154.078712280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.931201 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:25 crc kubenswrapper[4771]: E1002 09:39:25.931569 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:26.431553508 +0000 UTC m=+154.079238575 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.981290 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:25 crc kubenswrapper[4771]: I1002 09:39:25.981785 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.041033 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.047711 4771 patch_prober.go:28] interesting pod/apiserver-76f77b778f-sv9ls container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 02 09:39:26 crc kubenswrapper[4771]: [+]log ok Oct 02 09:39:26 crc kubenswrapper[4771]: [+]etcd ok Oct 02 09:39:26 crc kubenswrapper[4771]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 02 09:39:26 crc kubenswrapper[4771]: [+]poststarthook/generic-apiserver-start-informers ok Oct 02 09:39:26 crc kubenswrapper[4771]: [+]poststarthook/max-in-flight-filter ok Oct 02 09:39:26 crc kubenswrapper[4771]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 02 09:39:26 crc kubenswrapper[4771]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 02 09:39:26 crc kubenswrapper[4771]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 02 09:39:26 crc kubenswrapper[4771]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 02 09:39:26 crc kubenswrapper[4771]: [+]poststarthook/project.openshift.io-projectcache ok Oct 02 09:39:26 crc kubenswrapper[4771]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 02 09:39:26 crc kubenswrapper[4771]: [+]poststarthook/openshift.io-startinformers ok Oct 02 09:39:26 crc kubenswrapper[4771]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 02 09:39:26 crc kubenswrapper[4771]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 02 09:39:26 crc kubenswrapper[4771]: livez check failed Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.047806 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" podUID="961e712e-5760-4436-9be7-6bb6be2f8cf3" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:39:26 crc kubenswrapper[4771]: E1002 09:39:26.053786 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:26.543103462 +0000 UTC m=+154.190788529 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.071521 4771 patch_prober.go:28] interesting pod/router-default-5444994796-z7f8f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:39:26 crc kubenswrapper[4771]: [-]has-synced failed: reason withheld Oct 02 09:39:26 crc kubenswrapper[4771]: [+]process-running ok Oct 02 09:39:26 crc kubenswrapper[4771]: healthz check failed Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.071620 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z7f8f" podUID="12fbfa54-08b0-41aa-ba11-80f84d2724cf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.077581 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.085016 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-8dj62" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.094229 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqp4t" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.097089 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qd4wr"] Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.098899 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.110829 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.113507 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.111684 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.113818 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.119819 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.126805 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qd4wr"] Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.144226 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:26 crc kubenswrapper[4771]: E1002 09:39:26.144944 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:26.644926475 +0000 UTC m=+154.292611542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.257766 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.258157 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/820cd711-01de-4cb9-8dac-9cf33c165deb-utilities\") pod \"redhat-marketplace-qd4wr\" (UID: \"820cd711-01de-4cb9-8dac-9cf33c165deb\") " pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.258201 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvtm7\" (UniqueName: \"kubernetes.io/projected/820cd711-01de-4cb9-8dac-9cf33c165deb-kube-api-access-qvtm7\") pod \"redhat-marketplace-qd4wr\" (UID: \"820cd711-01de-4cb9-8dac-9cf33c165deb\") " pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.258232 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/820cd711-01de-4cb9-8dac-9cf33c165deb-catalog-content\") pod \"redhat-marketplace-qd4wr\" (UID: \"820cd711-01de-4cb9-8dac-9cf33c165deb\") " pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:39:26 crc kubenswrapper[4771]: E1002 09:39:26.258358 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:26.758314392 +0000 UTC m=+154.405999479 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.359096 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/820cd711-01de-4cb9-8dac-9cf33c165deb-utilities\") pod \"redhat-marketplace-qd4wr\" (UID: \"820cd711-01de-4cb9-8dac-9cf33c165deb\") " pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.359185 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvtm7\" (UniqueName: \"kubernetes.io/projected/820cd711-01de-4cb9-8dac-9cf33c165deb-kube-api-access-qvtm7\") pod \"redhat-marketplace-qd4wr\" (UID: \"820cd711-01de-4cb9-8dac-9cf33c165deb\") " pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.359229 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/820cd711-01de-4cb9-8dac-9cf33c165deb-catalog-content\") pod \"redhat-marketplace-qd4wr\" (UID: \"820cd711-01de-4cb9-8dac-9cf33c165deb\") " pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.359270 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:26 crc kubenswrapper[4771]: E1002 09:39:26.359603 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:26.859589399 +0000 UTC m=+154.507274466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.360254 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/820cd711-01de-4cb9-8dac-9cf33c165deb-utilities\") pod \"redhat-marketplace-qd4wr\" (UID: \"820cd711-01de-4cb9-8dac-9cf33c165deb\") " pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.360851 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/820cd711-01de-4cb9-8dac-9cf33c165deb-catalog-content\") pod \"redhat-marketplace-qd4wr\" (UID: \"820cd711-01de-4cb9-8dac-9cf33c165deb\") " pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.378822 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvtm7\" (UniqueName: \"kubernetes.io/projected/820cd711-01de-4cb9-8dac-9cf33c165deb-kube-api-access-qvtm7\") pod \"redhat-marketplace-qd4wr\" (UID: \"820cd711-01de-4cb9-8dac-9cf33c165deb\") " pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.460148 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:26 crc kubenswrapper[4771]: E1002 09:39:26.460437 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:26.960378213 +0000 UTC m=+154.608063300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.460784 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:26 crc kubenswrapper[4771]: E1002 09:39:26.461169 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:26.961154165 +0000 UTC m=+154.608839232 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.482847 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.487327 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2r4wz"] Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.488529 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:39:26 crc kubenswrapper[4771]: W1002 09:39:26.492989 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda072c8af_e1fb_4af2_98ba_e023152b4cfd.slice/crio-07586be142602c4b17a1a4c4be2df2f9a5aa6d7289d7c91811e0c40d47618211 WatchSource:0}: Error finding container 07586be142602c4b17a1a4c4be2df2f9a5aa6d7289d7c91811e0c40d47618211: Status 404 returned error can't find the container with id 07586be142602c4b17a1a4c4be2df2f9a5aa6d7289d7c91811e0c40d47618211 Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.508322 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2r4wz"] Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.566448 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:26 crc kubenswrapper[4771]: E1002 09:39:26.566834 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:27.066809139 +0000 UTC m=+154.714494206 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.567006 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-utilities\") pod \"redhat-marketplace-2r4wz\" (UID: \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\") " pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.567163 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.567303 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-catalog-content\") pod \"redhat-marketplace-2r4wz\" (UID: \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\") " pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.567347 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlr7b\" (UniqueName: \"kubernetes.io/projected/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-kube-api-access-zlr7b\") pod \"redhat-marketplace-2r4wz\" (UID: \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\") " pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:39:26 crc kubenswrapper[4771]: E1002 09:39:26.567561 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:27.06754911 +0000 UTC m=+154.715234347 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.596551 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.668631 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:26 crc kubenswrapper[4771]: E1002 09:39:26.668871 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:27.168848118 +0000 UTC m=+154.816533185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.668904 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-catalog-content\") pod \"redhat-marketplace-2r4wz\" (UID: \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\") " pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.668958 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlr7b\" (UniqueName: \"kubernetes.io/projected/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-kube-api-access-zlr7b\") pod \"redhat-marketplace-2r4wz\" (UID: \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\") " pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.669054 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-utilities\") pod \"redhat-marketplace-2r4wz\" (UID: \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\") " pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.670226 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-catalog-content\") pod \"redhat-marketplace-2r4wz\" (UID: \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\") " pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.670341 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-utilities\") pod \"redhat-marketplace-2r4wz\" (UID: \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\") " pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.689996 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlr7b\" (UniqueName: \"kubernetes.io/projected/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-kube-api-access-zlr7b\") pod \"redhat-marketplace-2r4wz\" (UID: \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\") " pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.749453 4771 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.757809 4771 generic.go:334] "Generic (PLEG): container finished" podID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" containerID="84897f45d59677d6d7babd051592e7e97f5a30e7b3c51ce469c748e6358b6638" exitCode=0 Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.758198 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l5tmt" event={"ID":"8fcbdb41-18a3-4977-aa8b-28f5af63576f","Type":"ContainerDied","Data":"84897f45d59677d6d7babd051592e7e97f5a30e7b3c51ce469c748e6358b6638"} Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.760642 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.762895 4771 generic.go:334] "Generic (PLEG): container finished" podID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" containerID="f4658ab7ce7bd35401056ca00aa46e24cc8ca45c958253cf26a12ecc5f168935" exitCode=0 Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.763073 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2gjk" event={"ID":"d01d796c-0c02-4c8d-b423-e32fc4b187d6","Type":"ContainerDied","Data":"f4658ab7ce7bd35401056ca00aa46e24cc8ca45c958253cf26a12ecc5f168935"} Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.772229 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:26 crc kubenswrapper[4771]: E1002 09:39:26.772591 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:27.272575876 +0000 UTC m=+154.920260943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.788043 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" event={"ID":"e3849bee-5c76-4308-8476-8abc1914d609","Type":"ContainerStarted","Data":"01b4d242bc08232bf115cad10d64c278385baecd103e531196127e436d7f5eec"} Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.799980 4771 generic.go:334] "Generic (PLEG): container finished" podID="d5107994-4460-4ac0-b975-6ba0fc73a542" containerID="bbb7cb32ebcf26c4b1eb082ebc2dc252e8381bb02aa7b849f8d33eb9c1a85ff3" exitCode=0 Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.800096 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zxgdb" event={"ID":"d5107994-4460-4ac0-b975-6ba0fc73a542","Type":"ContainerDied","Data":"bbb7cb32ebcf26c4b1eb082ebc2dc252e8381bb02aa7b849f8d33eb9c1a85ff3"} Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.800230 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zxgdb" event={"ID":"d5107994-4460-4ac0-b975-6ba0fc73a542","Type":"ContainerStarted","Data":"734974b790d0fe0cb141a1fd2d5b9b5fb6860704908bb0335c826be8af958848"} Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.808591 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.809931 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.811471 4771 generic.go:334] "Generic (PLEG): container finished" podID="d198e457-892f-40a9-9c39-46e4e67b79a8" containerID="3dcd486e2cff2498e972a9c310d3089828f49f0737d6deeaff7481bbf979c70a" exitCode=0 Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.812138 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lmpwh" event={"ID":"d198e457-892f-40a9-9c39-46e4e67b79a8","Type":"ContainerDied","Data":"3dcd486e2cff2498e972a9c310d3089828f49f0737d6deeaff7481bbf979c70a"} Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.815054 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.815358 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.828946 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.831689 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a072c8af-e1fb-4af2-98ba-e023152b4cfd","Type":"ContainerStarted","Data":"07586be142602c4b17a1a4c4be2df2f9a5aa6d7289d7c91811e0c40d47618211"} Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.862178 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.884733 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:26 crc kubenswrapper[4771]: E1002 09:39:26.899603 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:27.399557866 +0000 UTC m=+155.047242933 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.900765 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:26 crc kubenswrapper[4771]: E1002 09:39:26.903076 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:39:27.403034907 +0000 UTC m=+155.050720164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p2srt" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:26 crc kubenswrapper[4771]: I1002 09:39:26.910729 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qd4wr"] Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.002528 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.002980 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/def237c1-6f8f-40a2-9533-8854ba5079d5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"def237c1-6f8f-40a2-9533-8854ba5079d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.003019 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/def237c1-6f8f-40a2-9533-8854ba5079d5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"def237c1-6f8f-40a2-9533-8854ba5079d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:39:27 crc kubenswrapper[4771]: E1002 09:39:27.005058 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:39:27.504998284 +0000 UTC m=+155.152683491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.014719 4771 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-02T09:39:26.749763797Z","Handler":null,"Name":""} Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.024277 4771 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.024311 4771 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.035493 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.058368 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.064777 4771 patch_prober.go:28] interesting pod/router-default-5444994796-z7f8f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:39:27 crc kubenswrapper[4771]: [-]has-synced failed: reason withheld Oct 02 09:39:27 crc kubenswrapper[4771]: [+]process-running ok Oct 02 09:39:27 crc kubenswrapper[4771]: healthz check failed Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.064825 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z7f8f" podUID="12fbfa54-08b0-41aa-ba11-80f84d2724cf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.104218 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.104319 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/def237c1-6f8f-40a2-9533-8854ba5079d5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"def237c1-6f8f-40a2-9533-8854ba5079d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.104343 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/def237c1-6f8f-40a2-9533-8854ba5079d5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"def237c1-6f8f-40a2-9533-8854ba5079d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.104604 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/def237c1-6f8f-40a2-9533-8854ba5079d5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"def237c1-6f8f-40a2-9533-8854ba5079d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.128350 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/def237c1-6f8f-40a2-9533-8854ba5079d5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"def237c1-6f8f-40a2-9533-8854ba5079d5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.133564 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2r4wz"] Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.146293 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.245999 4771 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.246089 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.298315 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p2srt\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.298426 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d75tj"] Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.299968 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.302466 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.311459 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.311632 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e959f55-3b26-412b-9128-b9c26fd835b5-utilities\") pod \"redhat-operators-d75tj\" (UID: \"2e959f55-3b26-412b-9128-b9c26fd835b5\") " pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.311699 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e959f55-3b26-412b-9128-b9c26fd835b5-catalog-content\") pod \"redhat-operators-d75tj\" (UID: \"2e959f55-3b26-412b-9128-b9c26fd835b5\") " pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.311799 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qrmx\" (UniqueName: \"kubernetes.io/projected/2e959f55-3b26-412b-9128-b9c26fd835b5-kube-api-access-5qrmx\") pod \"redhat-operators-d75tj\" (UID: \"2e959f55-3b26-412b-9128-b9c26fd835b5\") " pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.313176 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d75tj"] Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.318680 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.399813 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.414176 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e959f55-3b26-412b-9128-b9c26fd835b5-utilities\") pod \"redhat-operators-d75tj\" (UID: \"2e959f55-3b26-412b-9128-b9c26fd835b5\") " pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.414235 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e959f55-3b26-412b-9128-b9c26fd835b5-catalog-content\") pod \"redhat-operators-d75tj\" (UID: \"2e959f55-3b26-412b-9128-b9c26fd835b5\") " pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.414276 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qrmx\" (UniqueName: \"kubernetes.io/projected/2e959f55-3b26-412b-9128-b9c26fd835b5-kube-api-access-5qrmx\") pod \"redhat-operators-d75tj\" (UID: \"2e959f55-3b26-412b-9128-b9c26fd835b5\") " pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.415082 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e959f55-3b26-412b-9128-b9c26fd835b5-utilities\") pod \"redhat-operators-d75tj\" (UID: \"2e959f55-3b26-412b-9128-b9c26fd835b5\") " pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.415330 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e959f55-3b26-412b-9128-b9c26fd835b5-catalog-content\") pod \"redhat-operators-d75tj\" (UID: \"2e959f55-3b26-412b-9128-b9c26fd835b5\") " pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.464013 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qrmx\" (UniqueName: \"kubernetes.io/projected/2e959f55-3b26-412b-9128-b9c26fd835b5-kube-api-access-5qrmx\") pod \"redhat-operators-d75tj\" (UID: \"2e959f55-3b26-412b-9128-b9c26fd835b5\") " pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.503556 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.628624 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p2srt"] Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.642400 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:39:27 crc kubenswrapper[4771]: W1002 09:39:27.656092 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod675e6316_9501_4ac0_a134_de142b7bfcd4.slice/crio-a8d14d1a87916d0224041f958acc63022f79c22db3c8500c83af3329ec4b250c WatchSource:0}: Error finding container a8d14d1a87916d0224041f958acc63022f79c22db3c8500c83af3329ec4b250c: Status 404 returned error can't find the container with id a8d14d1a87916d0224041f958acc63022f79c22db3c8500c83af3329ec4b250c Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.701476 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.702230 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z4dmh"] Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.704158 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z4dmh"] Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.704279 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.820967 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af193d2e-b44e-4519-84c3-cd40eb9eae02-utilities\") pod \"redhat-operators-z4dmh\" (UID: \"af193d2e-b44e-4519-84c3-cd40eb9eae02\") " pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.821493 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af193d2e-b44e-4519-84c3-cd40eb9eae02-catalog-content\") pod \"redhat-operators-z4dmh\" (UID: \"af193d2e-b44e-4519-84c3-cd40eb9eae02\") " pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.821513 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6vkp\" (UniqueName: \"kubernetes.io/projected/af193d2e-b44e-4519-84c3-cd40eb9eae02-kube-api-access-k6vkp\") pod \"redhat-operators-z4dmh\" (UID: \"af193d2e-b44e-4519-84c3-cd40eb9eae02\") " pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.861963 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"def237c1-6f8f-40a2-9533-8854ba5079d5","Type":"ContainerStarted","Data":"4d84ceecc733d462037e9c6f6c7c131e13808584ae8c97097af902ae38887548"} Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.867608 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a072c8af-e1fb-4af2-98ba-e023152b4cfd","Type":"ContainerStarted","Data":"4254509cd679c8048891255fade35b1c12c510db4a6d244feed520783ff0b9e3"} Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.884120 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" event={"ID":"675e6316-9501-4ac0-a134-de142b7bfcd4","Type":"ContainerStarted","Data":"a8d14d1a87916d0224041f958acc63022f79c22db3c8500c83af3329ec4b250c"} Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.907057 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qd4wr" event={"ID":"820cd711-01de-4cb9-8dac-9cf33c165deb","Type":"ContainerStarted","Data":"7bbe45fcd3465f80cd94debd6a11af5b18fdb65ab3163cfcd9fc3f760be5f160"} Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.907102 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qd4wr" event={"ID":"820cd711-01de-4cb9-8dac-9cf33c165deb","Type":"ContainerStarted","Data":"462704a71879f2a7c351c6ea78fac04278f802ec87c1624b690ac685fab4da12"} Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.915821 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2r4wz" event={"ID":"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d","Type":"ContainerStarted","Data":"2c9947b9ff03254723557a9c0838a4e62747ae2168554de9c4aa45ed17f9573b"} Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.927892 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6vkp\" (UniqueName: \"kubernetes.io/projected/af193d2e-b44e-4519-84c3-cd40eb9eae02-kube-api-access-k6vkp\") pod \"redhat-operators-z4dmh\" (UID: \"af193d2e-b44e-4519-84c3-cd40eb9eae02\") " pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.928015 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af193d2e-b44e-4519-84c3-cd40eb9eae02-utilities\") pod \"redhat-operators-z4dmh\" (UID: \"af193d2e-b44e-4519-84c3-cd40eb9eae02\") " pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.928385 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af193d2e-b44e-4519-84c3-cd40eb9eae02-catalog-content\") pod \"redhat-operators-z4dmh\" (UID: \"af193d2e-b44e-4519-84c3-cd40eb9eae02\") " pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.929073 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af193d2e-b44e-4519-84c3-cd40eb9eae02-utilities\") pod \"redhat-operators-z4dmh\" (UID: \"af193d2e-b44e-4519-84c3-cd40eb9eae02\") " pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.931534 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af193d2e-b44e-4519-84c3-cd40eb9eae02-catalog-content\") pod \"redhat-operators-z4dmh\" (UID: \"af193d2e-b44e-4519-84c3-cd40eb9eae02\") " pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:39:27 crc kubenswrapper[4771]: I1002 09:39:27.931667 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d75tj"] Oct 02 09:39:28 crc kubenswrapper[4771]: I1002 09:39:28.061170 4771 patch_prober.go:28] interesting pod/router-default-5444994796-z7f8f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:39:28 crc kubenswrapper[4771]: [-]has-synced failed: reason withheld Oct 02 09:39:28 crc kubenswrapper[4771]: [+]process-running ok Oct 02 09:39:28 crc kubenswrapper[4771]: healthz check failed Oct 02 09:39:28 crc kubenswrapper[4771]: I1002 09:39:28.061284 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z7f8f" podUID="12fbfa54-08b0-41aa-ba11-80f84d2724cf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:39:28 crc kubenswrapper[4771]: I1002 09:39:28.065247 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6vkp\" (UniqueName: \"kubernetes.io/projected/af193d2e-b44e-4519-84c3-cd40eb9eae02-kube-api-access-k6vkp\") pod \"redhat-operators-z4dmh\" (UID: \"af193d2e-b44e-4519-84c3-cd40eb9eae02\") " pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:39:28 crc kubenswrapper[4771]: I1002 09:39:28.345420 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:39:28 crc kubenswrapper[4771]: I1002 09:39:28.524062 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z4dmh"] Oct 02 09:39:28 crc kubenswrapper[4771]: W1002 09:39:28.532818 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf193d2e_b44e_4519_84c3_cd40eb9eae02.slice/crio-8b0a1061f564bf37be3e8d2b8c5e5abc5bea512feaf5a48f43c43cc4b116f804 WatchSource:0}: Error finding container 8b0a1061f564bf37be3e8d2b8c5e5abc5bea512feaf5a48f43c43cc4b116f804: Status 404 returned error can't find the container with id 8b0a1061f564bf37be3e8d2b8c5e5abc5bea512feaf5a48f43c43cc4b116f804 Oct 02 09:39:28 crc kubenswrapper[4771]: I1002 09:39:28.924560 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d75tj" event={"ID":"2e959f55-3b26-412b-9128-b9c26fd835b5","Type":"ContainerStarted","Data":"9dd68d8d0319116e4c19b6d06a615202a61fe63516fafeeb6f2e536156615b0a"} Oct 02 09:39:28 crc kubenswrapper[4771]: I1002 09:39:28.928059 4771 generic.go:334] "Generic (PLEG): container finished" podID="820cd711-01de-4cb9-8dac-9cf33c165deb" containerID="7bbe45fcd3465f80cd94debd6a11af5b18fdb65ab3163cfcd9fc3f760be5f160" exitCode=0 Oct 02 09:39:28 crc kubenswrapper[4771]: I1002 09:39:28.928197 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qd4wr" event={"ID":"820cd711-01de-4cb9-8dac-9cf33c165deb","Type":"ContainerDied","Data":"7bbe45fcd3465f80cd94debd6a11af5b18fdb65ab3163cfcd9fc3f760be5f160"} Oct 02 09:39:28 crc kubenswrapper[4771]: I1002 09:39:28.932513 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" event={"ID":"e3849bee-5c76-4308-8476-8abc1914d609","Type":"ContainerStarted","Data":"30011917245f5293821a0c139c87699b712f904a4ffc06cdc85589f3f1f75f7e"} Oct 02 09:39:28 crc kubenswrapper[4771]: I1002 09:39:28.934101 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z4dmh" event={"ID":"af193d2e-b44e-4519-84c3-cd40eb9eae02","Type":"ContainerStarted","Data":"8b0a1061f564bf37be3e8d2b8c5e5abc5bea512feaf5a48f43c43cc4b116f804"} Oct 02 09:39:29 crc kubenswrapper[4771]: I1002 09:39:29.060933 4771 patch_prober.go:28] interesting pod/router-default-5444994796-z7f8f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:39:29 crc kubenswrapper[4771]: [-]has-synced failed: reason withheld Oct 02 09:39:29 crc kubenswrapper[4771]: [+]process-running ok Oct 02 09:39:29 crc kubenswrapper[4771]: healthz check failed Oct 02 09:39:29 crc kubenswrapper[4771]: I1002 09:39:29.060995 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z7f8f" podUID="12fbfa54-08b0-41aa-ba11-80f84d2724cf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:39:29 crc kubenswrapper[4771]: I1002 09:39:29.069984 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-n6z8c" Oct 02 09:39:29 crc kubenswrapper[4771]: I1002 09:39:29.940870 4771 generic.go:334] "Generic (PLEG): container finished" podID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" containerID="c624cb5c98fe498d6e7d9aa31abe6b73d3548fe0c047100744a5b67d920d655b" exitCode=0 Oct 02 09:39:29 crc kubenswrapper[4771]: I1002 09:39:29.940919 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2r4wz" event={"ID":"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d","Type":"ContainerDied","Data":"c624cb5c98fe498d6e7d9aa31abe6b73d3548fe0c047100744a5b67d920d655b"} Oct 02 09:39:30 crc kubenswrapper[4771]: I1002 09:39:30.059968 4771 patch_prober.go:28] interesting pod/router-default-5444994796-z7f8f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:39:30 crc kubenswrapper[4771]: [-]has-synced failed: reason withheld Oct 02 09:39:30 crc kubenswrapper[4771]: [+]process-running ok Oct 02 09:39:30 crc kubenswrapper[4771]: healthz check failed Oct 02 09:39:30 crc kubenswrapper[4771]: I1002 09:39:30.060030 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z7f8f" podUID="12fbfa54-08b0-41aa-ba11-80f84d2724cf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:39:30 crc kubenswrapper[4771]: I1002 09:39:30.986184 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:30 crc kubenswrapper[4771]: I1002 09:39:30.993180 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-sv9ls" Oct 02 09:39:31 crc kubenswrapper[4771]: I1002 09:39:31.065672 4771 patch_prober.go:28] interesting pod/router-default-5444994796-z7f8f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:39:31 crc kubenswrapper[4771]: [-]has-synced failed: reason withheld Oct 02 09:39:31 crc kubenswrapper[4771]: [+]process-running ok Oct 02 09:39:31 crc kubenswrapper[4771]: healthz check failed Oct 02 09:39:31 crc kubenswrapper[4771]: I1002 09:39:31.065755 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z7f8f" podUID="12fbfa54-08b0-41aa-ba11-80f84d2724cf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:39:32 crc kubenswrapper[4771]: I1002 09:39:32.060578 4771 patch_prober.go:28] interesting pod/router-default-5444994796-z7f8f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:39:32 crc kubenswrapper[4771]: [+]has-synced ok Oct 02 09:39:32 crc kubenswrapper[4771]: [+]process-running ok Oct 02 09:39:32 crc kubenswrapper[4771]: healthz check failed Oct 02 09:39:32 crc kubenswrapper[4771]: I1002 09:39:32.060684 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z7f8f" podUID="12fbfa54-08b0-41aa-ba11-80f84d2724cf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:39:32 crc kubenswrapper[4771]: I1002 09:39:32.981995 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-pr9z4" podStartSLOduration=19.981883662 podStartE2EDuration="19.981883662s" podCreationTimestamp="2025-10-02 09:39:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:32.979588696 +0000 UTC m=+160.627273783" watchObservedRunningTime="2025-10-02 09:39:32.981883662 +0000 UTC m=+160.629568729" Oct 02 09:39:33 crc kubenswrapper[4771]: I1002 09:39:33.065613 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:33 crc kubenswrapper[4771]: I1002 09:39:33.071085 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-z7f8f" Oct 02 09:39:33 crc kubenswrapper[4771]: I1002 09:39:33.964003 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"def237c1-6f8f-40a2-9533-8854ba5079d5","Type":"ContainerStarted","Data":"93eea854271ceebd238261656592ea7a18a0470f7096f15ec277b58f584488c2"} Oct 02 09:39:33 crc kubenswrapper[4771]: I1002 09:39:33.976357 4771 generic.go:334] "Generic (PLEG): container finished" podID="a072c8af-e1fb-4af2-98ba-e023152b4cfd" containerID="4254509cd679c8048891255fade35b1c12c510db4a6d244feed520783ff0b9e3" exitCode=0 Oct 02 09:39:33 crc kubenswrapper[4771]: I1002 09:39:33.976485 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a072c8af-e1fb-4af2-98ba-e023152b4cfd","Type":"ContainerDied","Data":"4254509cd679c8048891255fade35b1c12c510db4a6d244feed520783ff0b9e3"} Oct 02 09:39:33 crc kubenswrapper[4771]: I1002 09:39:33.986640 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=7.986621613 podStartE2EDuration="7.986621613s" podCreationTimestamp="2025-10-02 09:39:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:33.979446535 +0000 UTC m=+161.627131602" watchObservedRunningTime="2025-10-02 09:39:33.986621613 +0000 UTC m=+161.634306680" Oct 02 09:39:33 crc kubenswrapper[4771]: I1002 09:39:33.989784 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" event={"ID":"675e6316-9501-4ac0-a134-de142b7bfcd4","Type":"ContainerStarted","Data":"bcf430f2779768e996dae1998ca3fc09a6a43d5ff5eac647ab3f05b175ed0c5c"} Oct 02 09:39:33 crc kubenswrapper[4771]: I1002 09:39:33.989868 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:33 crc kubenswrapper[4771]: I1002 09:39:33.994400 4771 generic.go:334] "Generic (PLEG): container finished" podID="af193d2e-b44e-4519-84c3-cd40eb9eae02" containerID="458bf13535d90f38fe64922ba58f3d8ed071953bc24606003caebd6ade2cd1e9" exitCode=0 Oct 02 09:39:33 crc kubenswrapper[4771]: I1002 09:39:33.994900 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z4dmh" event={"ID":"af193d2e-b44e-4519-84c3-cd40eb9eae02","Type":"ContainerDied","Data":"458bf13535d90f38fe64922ba58f3d8ed071953bc24606003caebd6ade2cd1e9"} Oct 02 09:39:34 crc kubenswrapper[4771]: I1002 09:39:34.002435 4771 generic.go:334] "Generic (PLEG): container finished" podID="2e959f55-3b26-412b-9128-b9c26fd835b5" containerID="a3f4166383fe85e891dd2b644db2dc3b5a3bd16f720bb500a14f85d454837048" exitCode=0 Oct 02 09:39:34 crc kubenswrapper[4771]: I1002 09:39:34.002557 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d75tj" event={"ID":"2e959f55-3b26-412b-9128-b9c26fd835b5","Type":"ContainerDied","Data":"a3f4166383fe85e891dd2b644db2dc3b5a3bd16f720bb500a14f85d454837048"} Oct 02 09:39:34 crc kubenswrapper[4771]: I1002 09:39:34.014288 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" podStartSLOduration=136.014266242 podStartE2EDuration="2m16.014266242s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:34.013392907 +0000 UTC m=+161.661077974" watchObservedRunningTime="2025-10-02 09:39:34.014266242 +0000 UTC m=+161.661951309" Oct 02 09:39:35 crc kubenswrapper[4771]: I1002 09:39:35.013465 4771 generic.go:334] "Generic (PLEG): container finished" podID="def237c1-6f8f-40a2-9533-8854ba5079d5" containerID="93eea854271ceebd238261656592ea7a18a0470f7096f15ec277b58f584488c2" exitCode=0 Oct 02 09:39:35 crc kubenswrapper[4771]: I1002 09:39:35.015352 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"def237c1-6f8f-40a2-9533-8854ba5079d5","Type":"ContainerDied","Data":"93eea854271ceebd238261656592ea7a18a0470f7096f15ec277b58f584488c2"} Oct 02 09:39:35 crc kubenswrapper[4771]: I1002 09:39:35.364287 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:39:35 crc kubenswrapper[4771]: I1002 09:39:35.533384 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a072c8af-e1fb-4af2-98ba-e023152b4cfd-kubelet-dir\") pod \"a072c8af-e1fb-4af2-98ba-e023152b4cfd\" (UID: \"a072c8af-e1fb-4af2-98ba-e023152b4cfd\") " Oct 02 09:39:35 crc kubenswrapper[4771]: I1002 09:39:35.533517 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a072c8af-e1fb-4af2-98ba-e023152b4cfd-kube-api-access\") pod \"a072c8af-e1fb-4af2-98ba-e023152b4cfd\" (UID: \"a072c8af-e1fb-4af2-98ba-e023152b4cfd\") " Oct 02 09:39:35 crc kubenswrapper[4771]: I1002 09:39:35.534531 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a072c8af-e1fb-4af2-98ba-e023152b4cfd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a072c8af-e1fb-4af2-98ba-e023152b4cfd" (UID: "a072c8af-e1fb-4af2-98ba-e023152b4cfd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:39:35 crc kubenswrapper[4771]: I1002 09:39:35.553515 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a072c8af-e1fb-4af2-98ba-e023152b4cfd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a072c8af-e1fb-4af2-98ba-e023152b4cfd" (UID: "a072c8af-e1fb-4af2-98ba-e023152b4cfd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:39:35 crc kubenswrapper[4771]: I1002 09:39:35.635698 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a072c8af-e1fb-4af2-98ba-e023152b4cfd-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:35 crc kubenswrapper[4771]: I1002 09:39:35.635741 4771 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a072c8af-e1fb-4af2-98ba-e023152b4cfd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:35 crc kubenswrapper[4771]: I1002 09:39:35.883220 4771 patch_prober.go:28] interesting pod/console-f9d7485db-r664s container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 02 09:39:35 crc kubenswrapper[4771]: I1002 09:39:35.883286 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-r664s" podUID="67d2ef8d-5a0c-48fa-8233-9bb23e794da0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 02 09:39:36 crc kubenswrapper[4771]: I1002 09:39:36.034864 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a072c8af-e1fb-4af2-98ba-e023152b4cfd","Type":"ContainerDied","Data":"07586be142602c4b17a1a4c4be2df2f9a5aa6d7289d7c91811e0c40d47618211"} Oct 02 09:39:36 crc kubenswrapper[4771]: I1002 09:39:36.034908 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:39:36 crc kubenswrapper[4771]: I1002 09:39:36.034913 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07586be142602c4b17a1a4c4be2df2f9a5aa6d7289d7c91811e0c40d47618211" Oct 02 09:39:36 crc kubenswrapper[4771]: I1002 09:39:36.111447 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:39:36 crc kubenswrapper[4771]: I1002 09:39:36.111539 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:39:36 crc kubenswrapper[4771]: I1002 09:39:36.111560 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:39:36 crc kubenswrapper[4771]: I1002 09:39:36.111674 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:39:39 crc kubenswrapper[4771]: I1002 09:39:39.975758 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:39:40 crc kubenswrapper[4771]: I1002 09:39:40.057844 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"def237c1-6f8f-40a2-9533-8854ba5079d5","Type":"ContainerDied","Data":"4d84ceecc733d462037e9c6f6c7c131e13808584ae8c97097af902ae38887548"} Oct 02 09:39:40 crc kubenswrapper[4771]: I1002 09:39:40.057881 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d84ceecc733d462037e9c6f6c7c131e13808584ae8c97097af902ae38887548" Oct 02 09:39:40 crc kubenswrapper[4771]: I1002 09:39:40.057894 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:39:40 crc kubenswrapper[4771]: I1002 09:39:40.099423 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/def237c1-6f8f-40a2-9533-8854ba5079d5-kubelet-dir\") pod \"def237c1-6f8f-40a2-9533-8854ba5079d5\" (UID: \"def237c1-6f8f-40a2-9533-8854ba5079d5\") " Oct 02 09:39:40 crc kubenswrapper[4771]: I1002 09:39:40.099504 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/def237c1-6f8f-40a2-9533-8854ba5079d5-kube-api-access\") pod \"def237c1-6f8f-40a2-9533-8854ba5079d5\" (UID: \"def237c1-6f8f-40a2-9533-8854ba5079d5\") " Oct 02 09:39:40 crc kubenswrapper[4771]: I1002 09:39:40.099550 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/def237c1-6f8f-40a2-9533-8854ba5079d5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "def237c1-6f8f-40a2-9533-8854ba5079d5" (UID: "def237c1-6f8f-40a2-9533-8854ba5079d5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:39:40 crc kubenswrapper[4771]: I1002 09:39:40.099861 4771 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/def237c1-6f8f-40a2-9533-8854ba5079d5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:40 crc kubenswrapper[4771]: I1002 09:39:40.104749 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/def237c1-6f8f-40a2-9533-8854ba5079d5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "def237c1-6f8f-40a2-9533-8854ba5079d5" (UID: "def237c1-6f8f-40a2-9533-8854ba5079d5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:39:40 crc kubenswrapper[4771]: I1002 09:39:40.201451 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/def237c1-6f8f-40a2-9533-8854ba5079d5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:40 crc kubenswrapper[4771]: I1002 09:39:40.808602 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:39:40 crc kubenswrapper[4771]: I1002 09:39:40.827186 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5431d-4595-4b96-a8b8-8953e3dffb53-metrics-certs\") pod \"network-metrics-daemon-zp7kh\" (UID: \"20d5431d-4595-4b96-a8b8-8953e3dffb53\") " pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:39:41 crc kubenswrapper[4771]: I1002 09:39:41.004395 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zp7kh" Oct 02 09:39:42 crc kubenswrapper[4771]: I1002 09:39:42.146201 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:39:42 crc kubenswrapper[4771]: I1002 09:39:42.146275 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:39:45 crc kubenswrapper[4771]: I1002 09:39:45.881730 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:45 crc kubenswrapper[4771]: I1002 09:39:45.885600 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:39:46 crc kubenswrapper[4771]: I1002 09:39:46.110993 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:39:46 crc kubenswrapper[4771]: I1002 09:39:46.111409 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:39:46 crc kubenswrapper[4771]: I1002 09:39:46.111217 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:39:46 crc kubenswrapper[4771]: I1002 09:39:46.111543 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:39:46 crc kubenswrapper[4771]: I1002 09:39:46.111605 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-vk4pw" Oct 02 09:39:46 crc kubenswrapper[4771]: I1002 09:39:46.112359 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:39:46 crc kubenswrapper[4771]: I1002 09:39:46.112432 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:39:46 crc kubenswrapper[4771]: I1002 09:39:46.113697 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"79769cf9608f05c46ac87c4ce509d61f44810c565b8d8e1dd6d064e2c4ea7008"} pod="openshift-console/downloads-7954f5f757-vk4pw" containerMessage="Container download-server failed liveness probe, will be restarted" Oct 02 09:39:46 crc kubenswrapper[4771]: I1002 09:39:46.113828 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" containerID="cri-o://79769cf9608f05c46ac87c4ce509d61f44810c565b8d8e1dd6d064e2c4ea7008" gracePeriod=2 Oct 02 09:39:47 crc kubenswrapper[4771]: I1002 09:39:47.092940 4771 generic.go:334] "Generic (PLEG): container finished" podID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerID="79769cf9608f05c46ac87c4ce509d61f44810c565b8d8e1dd6d064e2c4ea7008" exitCode=0 Oct 02 09:39:47 crc kubenswrapper[4771]: I1002 09:39:47.092984 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vk4pw" event={"ID":"77e60bc8-1d3d-4465-8dea-d7840a681b32","Type":"ContainerDied","Data":"79769cf9608f05c46ac87c4ce509d61f44810c565b8d8e1dd6d064e2c4ea7008"} Oct 02 09:39:54 crc kubenswrapper[4771]: E1002 09:39:54.976668 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 09:39:54 crc kubenswrapper[4771]: E1002 09:39:54.977380 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8bpzm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-l5tmt_openshift-marketplace(8fcbdb41-18a3-4977-aa8b-28f5af63576f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:39:54 crc kubenswrapper[4771]: E1002 09:39:54.978644 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-l5tmt" podUID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" Oct 02 09:39:55 crc kubenswrapper[4771]: E1002 09:39:55.899103 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-l5tmt" podUID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" Oct 02 09:39:56 crc kubenswrapper[4771]: I1002 09:39:56.112976 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:39:56 crc kubenswrapper[4771]: I1002 09:39:56.113581 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:39:56 crc kubenswrapper[4771]: I1002 09:39:56.328155 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zp7kh"] Oct 02 09:39:56 crc kubenswrapper[4771]: W1002 09:39:56.336103 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20d5431d_4595_4b96_a8b8_8953e3dffb53.slice/crio-4bfaa41921aaea9468dc0cd0fcc68e8003424c2617b548d332691cdf2fb65920 WatchSource:0}: Error finding container 4bfaa41921aaea9468dc0cd0fcc68e8003424c2617b548d332691cdf2fb65920: Status 404 returned error can't find the container with id 4bfaa41921aaea9468dc0cd0fcc68e8003424c2617b548d332691cdf2fb65920 Oct 02 09:39:57 crc kubenswrapper[4771]: I1002 09:39:57.057844 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q2xlm" Oct 02 09:39:57 crc kubenswrapper[4771]: I1002 09:39:57.140998 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" event={"ID":"20d5431d-4595-4b96-a8b8-8953e3dffb53","Type":"ContainerStarted","Data":"4bfaa41921aaea9468dc0cd0fcc68e8003424c2617b548d332691cdf2fb65920"} Oct 02 09:39:57 crc kubenswrapper[4771]: I1002 09:39:57.407615 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:39:58 crc kubenswrapper[4771]: I1002 09:39:58.150677 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" event={"ID":"20d5431d-4595-4b96-a8b8-8953e3dffb53","Type":"ContainerStarted","Data":"2cc8b90fce8492e0022dbef4d111a13c3dc58f7fcbeb4c092b2a9b98d38b508a"} Oct 02 09:39:58 crc kubenswrapper[4771]: I1002 09:39:58.153644 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vk4pw" event={"ID":"77e60bc8-1d3d-4465-8dea-d7840a681b32","Type":"ContainerStarted","Data":"f23614b653152266463c380f9cd024cc5f6e3df999d5f1e437980b7c1bf18e90"} Oct 02 09:40:00 crc kubenswrapper[4771]: E1002 09:40:00.045979 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 09:40:00 crc kubenswrapper[4771]: E1002 09:40:00.046443 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bm85t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lmpwh_openshift-marketplace(d198e457-892f-40a9-9c39-46e4e67b79a8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:40:00 crc kubenswrapper[4771]: E1002 09:40:00.047629 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-lmpwh" podUID="d198e457-892f-40a9-9c39-46e4e67b79a8" Oct 02 09:40:00 crc kubenswrapper[4771]: I1002 09:40:00.170087 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:40:00 crc kubenswrapper[4771]: I1002 09:40:00.170166 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:40:00 crc kubenswrapper[4771]: E1002 09:40:00.171791 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lmpwh" podUID="d198e457-892f-40a9-9c39-46e4e67b79a8" Oct 02 09:40:00 crc kubenswrapper[4771]: E1002 09:40:00.390765 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 09:40:00 crc kubenswrapper[4771]: E1002 09:40:00.390960 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b5tgt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zxgdb_openshift-marketplace(d5107994-4460-4ac0-b975-6ba0fc73a542): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:40:00 crc kubenswrapper[4771]: E1002 09:40:00.392406 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zxgdb" podUID="d5107994-4460-4ac0-b975-6ba0fc73a542" Oct 02 09:40:00 crc kubenswrapper[4771]: E1002 09:40:00.992442 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 09:40:00 crc kubenswrapper[4771]: E1002 09:40:00.992641 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m47qg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-k2gjk_openshift-marketplace(d01d796c-0c02-4c8d-b423-e32fc4b187d6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:40:00 crc kubenswrapper[4771]: E1002 09:40:00.993762 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-k2gjk" podUID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" Oct 02 09:40:01 crc kubenswrapper[4771]: I1002 09:40:01.178170 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zp7kh" event={"ID":"20d5431d-4595-4b96-a8b8-8953e3dffb53","Type":"ContainerStarted","Data":"1bd6a4aba318fa39a932e47ca2bb1fa31ca3cfd6e0c8b0c8c331e254e4ba53f6"} Oct 02 09:40:01 crc kubenswrapper[4771]: I1002 09:40:01.221428 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-zp7kh" podStartSLOduration=163.22140567 podStartE2EDuration="2m43.22140567s" podCreationTimestamp="2025-10-02 09:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:40:01.218015802 +0000 UTC m=+188.865700869" watchObservedRunningTime="2025-10-02 09:40:01.22140567 +0000 UTC m=+188.869090737" Oct 02 09:40:02 crc kubenswrapper[4771]: E1002 09:40:02.683075 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\": context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 09:40:02 crc kubenswrapper[4771]: E1002 09:40:02.683366 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zlr7b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2r4wz_openshift-marketplace(ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\": context canceled" logger="UnhandledError" Oct 02 09:40:02 crc kubenswrapper[4771]: E1002 09:40:02.684706 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \\\"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\\\": context canceled\"" pod="openshift-marketplace/redhat-marketplace-2r4wz" podUID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" Oct 02 09:40:02 crc kubenswrapper[4771]: I1002 09:40:02.902602 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:40:05 crc kubenswrapper[4771]: E1002 09:40:05.816055 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zxgdb" podUID="d5107994-4460-4ac0-b975-6ba0fc73a542" Oct 02 09:40:05 crc kubenswrapper[4771]: E1002 09:40:05.816724 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-k2gjk" podUID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" Oct 02 09:40:06 crc kubenswrapper[4771]: I1002 09:40:06.111186 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-vk4pw" Oct 02 09:40:06 crc kubenswrapper[4771]: I1002 09:40:06.112928 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:40:06 crc kubenswrapper[4771]: I1002 09:40:06.112995 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:40:06 crc kubenswrapper[4771]: I1002 09:40:06.113034 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:40:06 crc kubenswrapper[4771]: I1002 09:40:06.113068 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:40:06 crc kubenswrapper[4771]: I1002 09:40:06.113300 4771 patch_prober.go:28] interesting pod/downloads-7954f5f757-vk4pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 02 09:40:06 crc kubenswrapper[4771]: I1002 09:40:06.113330 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vk4pw" podUID="77e60bc8-1d3d-4465-8dea-d7840a681b32" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 02 09:40:09 crc kubenswrapper[4771]: E1002 09:40:09.061619 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 09:40:09 crc kubenswrapper[4771]: E1002 09:40:09.062348 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5qrmx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-d75tj_openshift-marketplace(2e959f55-3b26-412b-9128-b9c26fd835b5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:40:09 crc kubenswrapper[4771]: E1002 09:40:09.064026 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-d75tj" podUID="2e959f55-3b26-412b-9128-b9c26fd835b5" Oct 02 09:40:09 crc kubenswrapper[4771]: E1002 09:40:09.221212 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-d75tj" podUID="2e959f55-3b26-412b-9128-b9c26fd835b5" Oct 02 09:40:10 crc kubenswrapper[4771]: E1002 09:40:10.718033 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\": context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 09:40:10 crc kubenswrapper[4771]: E1002 09:40:10.718362 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qvtm7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-qd4wr_openshift-marketplace(820cd711-01de-4cb9-8dac-9cf33c165deb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\": context canceled" logger="UnhandledError" Oct 02 09:40:10 crc kubenswrapper[4771]: E1002 09:40:10.719567 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \\\"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\\\": context canceled\"" pod="openshift-marketplace/redhat-marketplace-qd4wr" podUID="820cd711-01de-4cb9-8dac-9cf33c165deb" Oct 02 09:40:12 crc kubenswrapper[4771]: I1002 09:40:12.146558 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:40:12 crc kubenswrapper[4771]: I1002 09:40:12.147337 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:40:16 crc kubenswrapper[4771]: I1002 09:40:16.116889 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-vk4pw" Oct 02 09:40:41 crc kubenswrapper[4771]: I1002 09:40:41.385070 4771 generic.go:334] "Generic (PLEG): container finished" podID="af193d2e-b44e-4519-84c3-cd40eb9eae02" containerID="308e1018d51418fa8601fdbfccf059a17888d4431efb164a1770ff65efa2eaed" exitCode=0 Oct 02 09:40:41 crc kubenswrapper[4771]: I1002 09:40:41.385114 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z4dmh" event={"ID":"af193d2e-b44e-4519-84c3-cd40eb9eae02","Type":"ContainerDied","Data":"308e1018d51418fa8601fdbfccf059a17888d4431efb164a1770ff65efa2eaed"} Oct 02 09:40:41 crc kubenswrapper[4771]: I1002 09:40:41.389816 4771 generic.go:334] "Generic (PLEG): container finished" podID="d198e457-892f-40a9-9c39-46e4e67b79a8" containerID="6d36f3be8fed9b32e81e651936aa866379a702a39145fcc5d045753973d07baa" exitCode=0 Oct 02 09:40:41 crc kubenswrapper[4771]: I1002 09:40:41.389851 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lmpwh" event={"ID":"d198e457-892f-40a9-9c39-46e4e67b79a8","Type":"ContainerDied","Data":"6d36f3be8fed9b32e81e651936aa866379a702a39145fcc5d045753973d07baa"} Oct 02 09:40:41 crc kubenswrapper[4771]: I1002 09:40:41.395716 4771 generic.go:334] "Generic (PLEG): container finished" podID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" containerID="0f8692d0baddc9926fc14493c2eaa9a9675b5c943cd46bdf4a17dcdf2be5e31d" exitCode=0 Oct 02 09:40:41 crc kubenswrapper[4771]: I1002 09:40:41.395790 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l5tmt" event={"ID":"8fcbdb41-18a3-4977-aa8b-28f5af63576f","Type":"ContainerDied","Data":"0f8692d0baddc9926fc14493c2eaa9a9675b5c943cd46bdf4a17dcdf2be5e31d"} Oct 02 09:40:42 crc kubenswrapper[4771]: I1002 09:40:42.146329 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:40:42 crc kubenswrapper[4771]: I1002 09:40:42.146447 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:40:42 crc kubenswrapper[4771]: I1002 09:40:42.146533 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:40:42 crc kubenswrapper[4771]: I1002 09:40:42.150993 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:40:42 crc kubenswrapper[4771]: I1002 09:40:42.151169 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a" gracePeriod=600 Oct 02 09:40:42 crc kubenswrapper[4771]: I1002 09:40:42.411418 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a" exitCode=0 Oct 02 09:40:42 crc kubenswrapper[4771]: I1002 09:40:42.411504 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a"} Oct 02 09:40:43 crc kubenswrapper[4771]: I1002 09:40:43.431973 4771 generic.go:334] "Generic (PLEG): container finished" podID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" containerID="471320cc6a54823504db3ed78e0e46340b5c3c5d5b563e9b81fa854100ca4a62" exitCode=0 Oct 02 09:40:43 crc kubenswrapper[4771]: I1002 09:40:43.432062 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2gjk" event={"ID":"d01d796c-0c02-4c8d-b423-e32fc4b187d6","Type":"ContainerDied","Data":"471320cc6a54823504db3ed78e0e46340b5c3c5d5b563e9b81fa854100ca4a62"} Oct 02 09:40:43 crc kubenswrapper[4771]: I1002 09:40:43.440456 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z4dmh" event={"ID":"af193d2e-b44e-4519-84c3-cd40eb9eae02","Type":"ContainerStarted","Data":"f9285adc21f4759c3d7a1bfea6c948ee20f422529775c821c3f87852cd8bc693"} Oct 02 09:40:43 crc kubenswrapper[4771]: I1002 09:40:43.447450 4771 generic.go:334] "Generic (PLEG): container finished" podID="d5107994-4460-4ac0-b975-6ba0fc73a542" containerID="61b8091deae05909bcc3ee4ad9113966503061a50ef311fbca77c022394a9e8f" exitCode=0 Oct 02 09:40:43 crc kubenswrapper[4771]: I1002 09:40:43.447527 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zxgdb" event={"ID":"d5107994-4460-4ac0-b975-6ba0fc73a542","Type":"ContainerDied","Data":"61b8091deae05909bcc3ee4ad9113966503061a50ef311fbca77c022394a9e8f"} Oct 02 09:40:43 crc kubenswrapper[4771]: I1002 09:40:43.452626 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lmpwh" event={"ID":"d198e457-892f-40a9-9c39-46e4e67b79a8","Type":"ContainerStarted","Data":"48a320f73eb7c7c10577a2f7879c39ac20182fc4ac06befaa2c8436de189c3f4"} Oct 02 09:40:43 crc kubenswrapper[4771]: I1002 09:40:43.477378 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"78b7a8d589783481aba6ddc2e3b7cc7710dc210299af82522a169351aa5ee352"} Oct 02 09:40:43 crc kubenswrapper[4771]: I1002 09:40:43.480167 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lmpwh" podStartSLOduration=3.401649842 podStartE2EDuration="1m19.480142083s" podCreationTimestamp="2025-10-02 09:39:24 +0000 UTC" firstStartedPulling="2025-10-02 09:39:26.814098056 +0000 UTC m=+154.461783113" lastFinishedPulling="2025-10-02 09:40:42.892590287 +0000 UTC m=+230.540275354" observedRunningTime="2025-10-02 09:40:43.47470044 +0000 UTC m=+231.122385497" watchObservedRunningTime="2025-10-02 09:40:43.480142083 +0000 UTC m=+231.127827150" Oct 02 09:40:43 crc kubenswrapper[4771]: I1002 09:40:43.481840 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l5tmt" event={"ID":"8fcbdb41-18a3-4977-aa8b-28f5af63576f","Type":"ContainerStarted","Data":"9379c2b224c899e1c699864e501137b5c1b464c50a2ec55f83668abab770b06d"} Oct 02 09:40:43 crc kubenswrapper[4771]: I1002 09:40:43.492752 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z4dmh" podStartSLOduration=7.793090774 podStartE2EDuration="1m16.492735778s" podCreationTimestamp="2025-10-02 09:39:27 +0000 UTC" firstStartedPulling="2025-10-02 09:39:34.000306458 +0000 UTC m=+161.647991535" lastFinishedPulling="2025-10-02 09:40:42.699951472 +0000 UTC m=+230.347636539" observedRunningTime="2025-10-02 09:40:43.489573309 +0000 UTC m=+231.137258376" watchObservedRunningTime="2025-10-02 09:40:43.492735778 +0000 UTC m=+231.140420845" Oct 02 09:40:43 crc kubenswrapper[4771]: I1002 09:40:43.554992 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l5tmt" podStartSLOduration=4.058193807 podStartE2EDuration="1m19.554969484s" podCreationTimestamp="2025-10-02 09:39:24 +0000 UTC" firstStartedPulling="2025-10-02 09:39:26.760303091 +0000 UTC m=+154.407988158" lastFinishedPulling="2025-10-02 09:40:42.257078768 +0000 UTC m=+229.904763835" observedRunningTime="2025-10-02 09:40:43.552019851 +0000 UTC m=+231.199704918" watchObservedRunningTime="2025-10-02 09:40:43.554969484 +0000 UTC m=+231.202654551" Oct 02 09:40:44 crc kubenswrapper[4771]: I1002 09:40:44.494222 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2gjk" event={"ID":"d01d796c-0c02-4c8d-b423-e32fc4b187d6","Type":"ContainerStarted","Data":"17a1d470be830766fa2aefd7892f1a10f83326d0a3db8b7c0cbaf704fe3b3bc6"} Oct 02 09:40:44 crc kubenswrapper[4771]: I1002 09:40:44.503845 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zxgdb" event={"ID":"d5107994-4460-4ac0-b975-6ba0fc73a542","Type":"ContainerStarted","Data":"41fc38381818b0c2afea2cb7dc2902e2fab54fe810ae2c71b631ba9de38676da"} Oct 02 09:40:44 crc kubenswrapper[4771]: I1002 09:40:44.519959 4771 generic.go:334] "Generic (PLEG): container finished" podID="2e959f55-3b26-412b-9128-b9c26fd835b5" containerID="ea400f0bf9a2031cc4cb0b4168fe1dc30adeb92ef43f4418b4dba2dceaa205d0" exitCode=0 Oct 02 09:40:44 crc kubenswrapper[4771]: I1002 09:40:44.520080 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d75tj" event={"ID":"2e959f55-3b26-412b-9128-b9c26fd835b5","Type":"ContainerDied","Data":"ea400f0bf9a2031cc4cb0b4168fe1dc30adeb92ef43f4418b4dba2dceaa205d0"} Oct 02 09:40:44 crc kubenswrapper[4771]: I1002 09:40:44.522713 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k2gjk" podStartSLOduration=3.417280629 podStartE2EDuration="1m20.522686257s" podCreationTimestamp="2025-10-02 09:39:24 +0000 UTC" firstStartedPulling="2025-10-02 09:39:26.778232909 +0000 UTC m=+154.425917966" lastFinishedPulling="2025-10-02 09:40:43.883638527 +0000 UTC m=+231.531323594" observedRunningTime="2025-10-02 09:40:44.520036462 +0000 UTC m=+232.167721529" watchObservedRunningTime="2025-10-02 09:40:44.522686257 +0000 UTC m=+232.170371324" Oct 02 09:40:44 crc kubenswrapper[4771]: I1002 09:40:44.537711 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:40:44 crc kubenswrapper[4771]: I1002 09:40:44.538158 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:40:44 crc kubenswrapper[4771]: I1002 09:40:44.541329 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zxgdb" podStartSLOduration=3.419789219 podStartE2EDuration="1m20.541308922s" podCreationTimestamp="2025-10-02 09:39:24 +0000 UTC" firstStartedPulling="2025-10-02 09:39:26.806567748 +0000 UTC m=+154.454252825" lastFinishedPulling="2025-10-02 09:40:43.928087461 +0000 UTC m=+231.575772528" observedRunningTime="2025-10-02 09:40:44.538902304 +0000 UTC m=+232.186587371" watchObservedRunningTime="2025-10-02 09:40:44.541308922 +0000 UTC m=+232.188993989" Oct 02 09:40:44 crc kubenswrapper[4771]: I1002 09:40:44.637889 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:40:44 crc kubenswrapper[4771]: I1002 09:40:44.637950 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:40:44 crc kubenswrapper[4771]: I1002 09:40:44.857834 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:40:44 crc kubenswrapper[4771]: I1002 09:40:44.858465 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:40:45 crc kubenswrapper[4771]: I1002 09:40:45.044513 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:40:45 crc kubenswrapper[4771]: I1002 09:40:45.044565 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:40:45 crc kubenswrapper[4771]: I1002 09:40:45.531458 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d75tj" event={"ID":"2e959f55-3b26-412b-9128-b9c26fd835b5","Type":"ContainerStarted","Data":"1d329a63150ec9fd6a8518cfbe7406ec5b2f866c918675cda2e281c58533db94"} Oct 02 09:40:45 crc kubenswrapper[4771]: I1002 09:40:45.568878 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d75tj" podStartSLOduration=7.353526831 podStartE2EDuration="1m18.568861903s" podCreationTimestamp="2025-10-02 09:39:27 +0000 UTC" firstStartedPulling="2025-10-02 09:39:34.010115692 +0000 UTC m=+161.657800759" lastFinishedPulling="2025-10-02 09:40:45.225450764 +0000 UTC m=+232.873135831" observedRunningTime="2025-10-02 09:40:45.566141956 +0000 UTC m=+233.213827023" watchObservedRunningTime="2025-10-02 09:40:45.568861903 +0000 UTC m=+233.216546970" Oct 02 09:40:45 crc kubenswrapper[4771]: I1002 09:40:45.744325 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-lmpwh" podUID="d198e457-892f-40a9-9c39-46e4e67b79a8" containerName="registry-server" probeResult="failure" output=< Oct 02 09:40:45 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 09:40:45 crc kubenswrapper[4771]: > Oct 02 09:40:45 crc kubenswrapper[4771]: I1002 09:40:45.748479 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-l5tmt" podUID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" containerName="registry-server" probeResult="failure" output=< Oct 02 09:40:45 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 09:40:45 crc kubenswrapper[4771]: > Oct 02 09:40:45 crc kubenswrapper[4771]: I1002 09:40:45.919913 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-zxgdb" podUID="d5107994-4460-4ac0-b975-6ba0fc73a542" containerName="registry-server" probeResult="failure" output=< Oct 02 09:40:45 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 09:40:45 crc kubenswrapper[4771]: > Oct 02 09:40:46 crc kubenswrapper[4771]: I1002 09:40:46.113680 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-k2gjk" podUID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" containerName="registry-server" probeResult="failure" output=< Oct 02 09:40:46 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 09:40:46 crc kubenswrapper[4771]: > Oct 02 09:40:47 crc kubenswrapper[4771]: I1002 09:40:47.644494 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:40:47 crc kubenswrapper[4771]: I1002 09:40:47.644536 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:40:48 crc kubenswrapper[4771]: I1002 09:40:48.346229 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:40:48 crc kubenswrapper[4771]: I1002 09:40:48.346576 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:40:48 crc kubenswrapper[4771]: I1002 09:40:48.436053 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:40:48 crc kubenswrapper[4771]: I1002 09:40:48.587351 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:40:48 crc kubenswrapper[4771]: I1002 09:40:48.692725 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d75tj" podUID="2e959f55-3b26-412b-9128-b9c26fd835b5" containerName="registry-server" probeResult="failure" output=< Oct 02 09:40:48 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 09:40:48 crc kubenswrapper[4771]: > Oct 02 09:40:49 crc kubenswrapper[4771]: I1002 09:40:49.621687 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z4dmh"] Oct 02 09:40:50 crc kubenswrapper[4771]: I1002 09:40:50.555932 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z4dmh" podUID="af193d2e-b44e-4519-84c3-cd40eb9eae02" containerName="registry-server" containerID="cri-o://f9285adc21f4759c3d7a1bfea6c948ee20f422529775c821c3f87852cd8bc693" gracePeriod=2 Oct 02 09:40:53 crc kubenswrapper[4771]: I1002 09:40:53.580443 4771 generic.go:334] "Generic (PLEG): container finished" podID="af193d2e-b44e-4519-84c3-cd40eb9eae02" containerID="f9285adc21f4759c3d7a1bfea6c948ee20f422529775c821c3f87852cd8bc693" exitCode=0 Oct 02 09:40:53 crc kubenswrapper[4771]: I1002 09:40:53.580517 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z4dmh" event={"ID":"af193d2e-b44e-4519-84c3-cd40eb9eae02","Type":"ContainerDied","Data":"f9285adc21f4759c3d7a1bfea6c948ee20f422529775c821c3f87852cd8bc693"} Oct 02 09:40:54 crc kubenswrapper[4771]: I1002 09:40:54.578277 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:40:54 crc kubenswrapper[4771]: I1002 09:40:54.615892 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:40:54 crc kubenswrapper[4771]: I1002 09:40:54.763476 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:40:54 crc kubenswrapper[4771]: I1002 09:40:54.807905 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:40:54 crc kubenswrapper[4771]: I1002 09:40:54.891304 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:40:54 crc kubenswrapper[4771]: I1002 09:40:54.926912 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:40:55 crc kubenswrapper[4771]: I1002 09:40:55.080638 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:40:55 crc kubenswrapper[4771]: I1002 09:40:55.116726 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:40:56 crc kubenswrapper[4771]: I1002 09:40:56.811561 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zxgdb"] Oct 02 09:40:56 crc kubenswrapper[4771]: I1002 09:40:56.811843 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zxgdb" podUID="d5107994-4460-4ac0-b975-6ba0fc73a542" containerName="registry-server" containerID="cri-o://41fc38381818b0c2afea2cb7dc2902e2fab54fe810ae2c71b631ba9de38676da" gracePeriod=2 Oct 02 09:40:57 crc kubenswrapper[4771]: I1002 09:40:57.009180 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k2gjk"] Oct 02 09:40:57 crc kubenswrapper[4771]: I1002 09:40:57.009482 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k2gjk" podUID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" containerName="registry-server" containerID="cri-o://17a1d470be830766fa2aefd7892f1a10f83326d0a3db8b7c0cbaf704fe3b3bc6" gracePeriod=2 Oct 02 09:40:57 crc kubenswrapper[4771]: I1002 09:40:57.680420 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:40:57 crc kubenswrapper[4771]: I1002 09:40:57.734462 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:40:58 crc kubenswrapper[4771]: E1002 09:40:58.346811 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f9285adc21f4759c3d7a1bfea6c948ee20f422529775c821c3f87852cd8bc693 is running failed: container process not found" containerID="f9285adc21f4759c3d7a1bfea6c948ee20f422529775c821c3f87852cd8bc693" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 09:40:58 crc kubenswrapper[4771]: E1002 09:40:58.347466 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f9285adc21f4759c3d7a1bfea6c948ee20f422529775c821c3f87852cd8bc693 is running failed: container process not found" containerID="f9285adc21f4759c3d7a1bfea6c948ee20f422529775c821c3f87852cd8bc693" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 09:40:58 crc kubenswrapper[4771]: E1002 09:40:58.347859 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f9285adc21f4759c3d7a1bfea6c948ee20f422529775c821c3f87852cd8bc693 is running failed: container process not found" containerID="f9285adc21f4759c3d7a1bfea6c948ee20f422529775c821c3f87852cd8bc693" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 09:40:58 crc kubenswrapper[4771]: E1002 09:40:58.347896 4771 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f9285adc21f4759c3d7a1bfea6c948ee20f422529775c821c3f87852cd8bc693 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-z4dmh" podUID="af193d2e-b44e-4519-84c3-cd40eb9eae02" containerName="registry-server" Oct 02 09:40:59 crc kubenswrapper[4771]: I1002 09:40:59.613039 4771 generic.go:334] "Generic (PLEG): container finished" podID="d5107994-4460-4ac0-b975-6ba0fc73a542" containerID="41fc38381818b0c2afea2cb7dc2902e2fab54fe810ae2c71b631ba9de38676da" exitCode=0 Oct 02 09:40:59 crc kubenswrapper[4771]: I1002 09:40:59.613103 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zxgdb" event={"ID":"d5107994-4460-4ac0-b975-6ba0fc73a542","Type":"ContainerDied","Data":"41fc38381818b0c2afea2cb7dc2902e2fab54fe810ae2c71b631ba9de38676da"} Oct 02 09:40:59 crc kubenswrapper[4771]: I1002 09:40:59.615281 4771 generic.go:334] "Generic (PLEG): container finished" podID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" containerID="17a1d470be830766fa2aefd7892f1a10f83326d0a3db8b7c0cbaf704fe3b3bc6" exitCode=0 Oct 02 09:40:59 crc kubenswrapper[4771]: I1002 09:40:59.615340 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2gjk" event={"ID":"d01d796c-0c02-4c8d-b423-e32fc4b187d6","Type":"ContainerDied","Data":"17a1d470be830766fa2aefd7892f1a10f83326d0a3db8b7c0cbaf704fe3b3bc6"} Oct 02 09:41:01 crc kubenswrapper[4771]: I1002 09:41:01.557774 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:41:01 crc kubenswrapper[4771]: I1002 09:41:01.624927 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z4dmh" event={"ID":"af193d2e-b44e-4519-84c3-cd40eb9eae02","Type":"ContainerDied","Data":"8b0a1061f564bf37be3e8d2b8c5e5abc5bea512feaf5a48f43c43cc4b116f804"} Oct 02 09:41:01 crc kubenswrapper[4771]: I1002 09:41:01.624966 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z4dmh" Oct 02 09:41:01 crc kubenswrapper[4771]: I1002 09:41:01.624977 4771 scope.go:117] "RemoveContainer" containerID="f9285adc21f4759c3d7a1bfea6c948ee20f422529775c821c3f87852cd8bc693" Oct 02 09:41:01 crc kubenswrapper[4771]: I1002 09:41:01.680360 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af193d2e-b44e-4519-84c3-cd40eb9eae02-utilities\") pod \"af193d2e-b44e-4519-84c3-cd40eb9eae02\" (UID: \"af193d2e-b44e-4519-84c3-cd40eb9eae02\") " Oct 02 09:41:01 crc kubenswrapper[4771]: I1002 09:41:01.680409 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af193d2e-b44e-4519-84c3-cd40eb9eae02-catalog-content\") pod \"af193d2e-b44e-4519-84c3-cd40eb9eae02\" (UID: \"af193d2e-b44e-4519-84c3-cd40eb9eae02\") " Oct 02 09:41:01 crc kubenswrapper[4771]: I1002 09:41:01.680435 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6vkp\" (UniqueName: \"kubernetes.io/projected/af193d2e-b44e-4519-84c3-cd40eb9eae02-kube-api-access-k6vkp\") pod \"af193d2e-b44e-4519-84c3-cd40eb9eae02\" (UID: \"af193d2e-b44e-4519-84c3-cd40eb9eae02\") " Oct 02 09:41:01 crc kubenswrapper[4771]: I1002 09:41:01.682029 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af193d2e-b44e-4519-84c3-cd40eb9eae02-utilities" (OuterVolumeSpecName: "utilities") pod "af193d2e-b44e-4519-84c3-cd40eb9eae02" (UID: "af193d2e-b44e-4519-84c3-cd40eb9eae02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:01 crc kubenswrapper[4771]: I1002 09:41:01.687117 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af193d2e-b44e-4519-84c3-cd40eb9eae02-kube-api-access-k6vkp" (OuterVolumeSpecName: "kube-api-access-k6vkp") pod "af193d2e-b44e-4519-84c3-cd40eb9eae02" (UID: "af193d2e-b44e-4519-84c3-cd40eb9eae02"). InnerVolumeSpecName "kube-api-access-k6vkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:41:01 crc kubenswrapper[4771]: I1002 09:41:01.781955 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af193d2e-b44e-4519-84c3-cd40eb9eae02-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:01 crc kubenswrapper[4771]: I1002 09:41:01.781995 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6vkp\" (UniqueName: \"kubernetes.io/projected/af193d2e-b44e-4519-84c3-cd40eb9eae02-kube-api-access-k6vkp\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:02 crc kubenswrapper[4771]: I1002 09:41:02.542696 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af193d2e-b44e-4519-84c3-cd40eb9eae02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af193d2e-b44e-4519-84c3-cd40eb9eae02" (UID: "af193d2e-b44e-4519-84c3-cd40eb9eae02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:02 crc kubenswrapper[4771]: I1002 09:41:02.593167 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af193d2e-b44e-4519-84c3-cd40eb9eae02-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:02 crc kubenswrapper[4771]: I1002 09:41:02.851881 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z4dmh"] Oct 02 09:41:02 crc kubenswrapper[4771]: I1002 09:41:02.854679 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z4dmh"] Oct 02 09:41:03 crc kubenswrapper[4771]: I1002 09:41:03.687783 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af193d2e-b44e-4519-84c3-cd40eb9eae02" path="/var/lib/kubelet/pods/af193d2e-b44e-4519-84c3-cd40eb9eae02/volumes" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.491191 4771 scope.go:117] "RemoveContainer" containerID="308e1018d51418fa8601fdbfccf059a17888d4431efb164a1770ff65efa2eaed" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.528140 4771 scope.go:117] "RemoveContainer" containerID="458bf13535d90f38fe64922ba58f3d8ed071953bc24606003caebd6ade2cd1e9" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.540653 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.565839 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.645438 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zxgdb" event={"ID":"d5107994-4460-4ac0-b975-6ba0fc73a542","Type":"ContainerDied","Data":"734974b790d0fe0cb141a1fd2d5b9b5fb6860704908bb0335c826be8af958848"} Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.645527 4771 scope.go:117] "RemoveContainer" containerID="41fc38381818b0c2afea2cb7dc2902e2fab54fe810ae2c71b631ba9de38676da" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.645460 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zxgdb" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.650714 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2gjk" event={"ID":"d01d796c-0c02-4c8d-b423-e32fc4b187d6","Type":"ContainerDied","Data":"15d56732786d0ff7d14712b3edce8421a246eddd28446e24712b343a510b79e0"} Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.650809 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2gjk" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.659916 4771 scope.go:117] "RemoveContainer" containerID="61b8091deae05909bcc3ee4ad9113966503061a50ef311fbca77c022394a9e8f" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.672255 4771 scope.go:117] "RemoveContainer" containerID="bbb7cb32ebcf26c4b1eb082ebc2dc252e8381bb02aa7b849f8d33eb9c1a85ff3" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.689605 4771 scope.go:117] "RemoveContainer" containerID="17a1d470be830766fa2aefd7892f1a10f83326d0a3db8b7c0cbaf704fe3b3bc6" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.701917 4771 scope.go:117] "RemoveContainer" containerID="471320cc6a54823504db3ed78e0e46340b5c3c5d5b563e9b81fa854100ca4a62" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.714301 4771 scope.go:117] "RemoveContainer" containerID="f4658ab7ce7bd35401056ca00aa46e24cc8ca45c958253cf26a12ecc5f168935" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.720392 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5107994-4460-4ac0-b975-6ba0fc73a542-catalog-content\") pod \"d5107994-4460-4ac0-b975-6ba0fc73a542\" (UID: \"d5107994-4460-4ac0-b975-6ba0fc73a542\") " Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.720461 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5107994-4460-4ac0-b975-6ba0fc73a542-utilities\") pod \"d5107994-4460-4ac0-b975-6ba0fc73a542\" (UID: \"d5107994-4460-4ac0-b975-6ba0fc73a542\") " Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.720528 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5tgt\" (UniqueName: \"kubernetes.io/projected/d5107994-4460-4ac0-b975-6ba0fc73a542-kube-api-access-b5tgt\") pod \"d5107994-4460-4ac0-b975-6ba0fc73a542\" (UID: \"d5107994-4460-4ac0-b975-6ba0fc73a542\") " Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.720570 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m47qg\" (UniqueName: \"kubernetes.io/projected/d01d796c-0c02-4c8d-b423-e32fc4b187d6-kube-api-access-m47qg\") pod \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\" (UID: \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\") " Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.720594 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d01d796c-0c02-4c8d-b423-e32fc4b187d6-catalog-content\") pod \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\" (UID: \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\") " Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.720647 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d01d796c-0c02-4c8d-b423-e32fc4b187d6-utilities\") pod \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\" (UID: \"d01d796c-0c02-4c8d-b423-e32fc4b187d6\") " Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.721336 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5107994-4460-4ac0-b975-6ba0fc73a542-utilities" (OuterVolumeSpecName: "utilities") pod "d5107994-4460-4ac0-b975-6ba0fc73a542" (UID: "d5107994-4460-4ac0-b975-6ba0fc73a542"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.721581 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d01d796c-0c02-4c8d-b423-e32fc4b187d6-utilities" (OuterVolumeSpecName: "utilities") pod "d01d796c-0c02-4c8d-b423-e32fc4b187d6" (UID: "d01d796c-0c02-4c8d-b423-e32fc4b187d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.726091 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5107994-4460-4ac0-b975-6ba0fc73a542-kube-api-access-b5tgt" (OuterVolumeSpecName: "kube-api-access-b5tgt") pod "d5107994-4460-4ac0-b975-6ba0fc73a542" (UID: "d5107994-4460-4ac0-b975-6ba0fc73a542"). InnerVolumeSpecName "kube-api-access-b5tgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.726963 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d01d796c-0c02-4c8d-b423-e32fc4b187d6-kube-api-access-m47qg" (OuterVolumeSpecName: "kube-api-access-m47qg") pod "d01d796c-0c02-4c8d-b423-e32fc4b187d6" (UID: "d01d796c-0c02-4c8d-b423-e32fc4b187d6"). InnerVolumeSpecName "kube-api-access-m47qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.821894 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5107994-4460-4ac0-b975-6ba0fc73a542-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.821929 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5tgt\" (UniqueName: \"kubernetes.io/projected/d5107994-4460-4ac0-b975-6ba0fc73a542-kube-api-access-b5tgt\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.821942 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m47qg\" (UniqueName: \"kubernetes.io/projected/d01d796c-0c02-4c8d-b423-e32fc4b187d6-kube-api-access-m47qg\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:04 crc kubenswrapper[4771]: I1002 09:41:04.821951 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d01d796c-0c02-4c8d-b423-e32fc4b187d6-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:05 crc kubenswrapper[4771]: I1002 09:41:05.185976 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5107994-4460-4ac0-b975-6ba0fc73a542-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5107994-4460-4ac0-b975-6ba0fc73a542" (UID: "d5107994-4460-4ac0-b975-6ba0fc73a542"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:05 crc kubenswrapper[4771]: I1002 09:41:05.228248 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5107994-4460-4ac0-b975-6ba0fc73a542-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:05 crc kubenswrapper[4771]: I1002 09:41:05.269985 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zxgdb"] Oct 02 09:41:05 crc kubenswrapper[4771]: I1002 09:41:05.274344 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zxgdb"] Oct 02 09:41:05 crc kubenswrapper[4771]: I1002 09:41:05.405022 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d01d796c-0c02-4c8d-b423-e32fc4b187d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d01d796c-0c02-4c8d-b423-e32fc4b187d6" (UID: "d01d796c-0c02-4c8d-b423-e32fc4b187d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:05 crc kubenswrapper[4771]: I1002 09:41:05.430661 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d01d796c-0c02-4c8d-b423-e32fc4b187d6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:05 crc kubenswrapper[4771]: I1002 09:41:05.579302 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k2gjk"] Oct 02 09:41:05 crc kubenswrapper[4771]: I1002 09:41:05.582297 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k2gjk"] Oct 02 09:41:05 crc kubenswrapper[4771]: I1002 09:41:05.688054 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" path="/var/lib/kubelet/pods/d01d796c-0c02-4c8d-b423-e32fc4b187d6/volumes" Oct 02 09:41:05 crc kubenswrapper[4771]: I1002 09:41:05.688806 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5107994-4460-4ac0-b975-6ba0fc73a542" path="/var/lib/kubelet/pods/d5107994-4460-4ac0-b975-6ba0fc73a542/volumes" Oct 02 09:41:07 crc kubenswrapper[4771]: E1002 09:41:07.180906 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 09:41:07 crc kubenswrapper[4771]: E1002 09:41:07.183380 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qvtm7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-qd4wr_openshift-marketplace(820cd711-01de-4cb9-8dac-9cf33c165deb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:41:07 crc kubenswrapper[4771]: E1002 09:41:07.184702 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-qd4wr" podUID="820cd711-01de-4cb9-8dac-9cf33c165deb" Oct 02 09:41:08 crc kubenswrapper[4771]: E1002 09:41:08.893770 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 09:41:08 crc kubenswrapper[4771]: E1002 09:41:08.893932 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zlr7b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2r4wz_openshift-marketplace(ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:41:08 crc kubenswrapper[4771]: E1002 09:41:08.895198 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-2r4wz" podUID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" Oct 02 09:41:17 crc kubenswrapper[4771]: E1002 09:41:17.683663 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-qd4wr" podUID="820cd711-01de-4cb9-8dac-9cf33c165deb" Oct 02 09:41:20 crc kubenswrapper[4771]: E1002 09:41:20.682538 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2r4wz" podUID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" Oct 02 09:41:31 crc kubenswrapper[4771]: I1002 09:41:31.811548 4771 generic.go:334] "Generic (PLEG): container finished" podID="820cd711-01de-4cb9-8dac-9cf33c165deb" containerID="e6389a097fc31f5a74c3d90d8d1ee7a1af07f6571bc0ea3c0090a1fa9ca2ddd1" exitCode=0 Oct 02 09:41:31 crc kubenswrapper[4771]: I1002 09:41:31.811702 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qd4wr" event={"ID":"820cd711-01de-4cb9-8dac-9cf33c165deb","Type":"ContainerDied","Data":"e6389a097fc31f5a74c3d90d8d1ee7a1af07f6571bc0ea3c0090a1fa9ca2ddd1"} Oct 02 09:41:32 crc kubenswrapper[4771]: I1002 09:41:32.826800 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qd4wr" event={"ID":"820cd711-01de-4cb9-8dac-9cf33c165deb","Type":"ContainerStarted","Data":"c16893f4f3f4fd1d2d368e83e1122bc898940a6c79df8ffbc0a63a8a0ed443e6"} Oct 02 09:41:32 crc kubenswrapper[4771]: I1002 09:41:32.846197 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qd4wr" podStartSLOduration=7.59208046 podStartE2EDuration="2m6.846180005s" podCreationTimestamp="2025-10-02 09:39:26 +0000 UTC" firstStartedPulling="2025-10-02 09:39:32.957617131 +0000 UTC m=+160.605302198" lastFinishedPulling="2025-10-02 09:41:32.211716676 +0000 UTC m=+279.859401743" observedRunningTime="2025-10-02 09:41:32.843813171 +0000 UTC m=+280.491498238" watchObservedRunningTime="2025-10-02 09:41:32.846180005 +0000 UTC m=+280.493865072" Oct 02 09:41:33 crc kubenswrapper[4771]: I1002 09:41:33.837687 4771 generic.go:334] "Generic (PLEG): container finished" podID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" containerID="9d24821999a730acde162c699127d7d5ae473270e9f51cd8afc12e1ce7036272" exitCode=0 Oct 02 09:41:33 crc kubenswrapper[4771]: I1002 09:41:33.837731 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2r4wz" event={"ID":"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d","Type":"ContainerDied","Data":"9d24821999a730acde162c699127d7d5ae473270e9f51cd8afc12e1ce7036272"} Oct 02 09:41:34 crc kubenswrapper[4771]: I1002 09:41:34.846397 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2r4wz" event={"ID":"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d","Type":"ContainerStarted","Data":"031ae6d60b824b867644fb346fda6f99ac137a96c5e80252d3c5fa69de0db797"} Oct 02 09:41:36 crc kubenswrapper[4771]: I1002 09:41:36.596962 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:41:36 crc kubenswrapper[4771]: I1002 09:41:36.597990 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:41:36 crc kubenswrapper[4771]: I1002 09:41:36.635910 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:41:36 crc kubenswrapper[4771]: I1002 09:41:36.651889 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2r4wz" podStartSLOduration=10.382870032 podStartE2EDuration="2m10.65186949s" podCreationTimestamp="2025-10-02 09:39:26 +0000 UTC" firstStartedPulling="2025-10-02 09:39:34.004939912 +0000 UTC m=+161.652624979" lastFinishedPulling="2025-10-02 09:41:34.27393937 +0000 UTC m=+281.921624437" observedRunningTime="2025-10-02 09:41:34.869214212 +0000 UTC m=+282.516899279" watchObservedRunningTime="2025-10-02 09:41:36.65186949 +0000 UTC m=+284.299554557" Oct 02 09:41:36 crc kubenswrapper[4771]: I1002 09:41:36.862918 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:41:36 crc kubenswrapper[4771]: I1002 09:41:36.862957 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:41:36 crc kubenswrapper[4771]: I1002 09:41:36.895966 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:41:37 crc kubenswrapper[4771]: I1002 09:41:37.921708 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:41:46 crc kubenswrapper[4771]: I1002 09:41:46.924099 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-gdzf9"] Oct 02 09:41:46 crc kubenswrapper[4771]: I1002 09:41:46.925445 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:41:47 crc kubenswrapper[4771]: I1002 09:41:47.032485 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2r4wz"] Oct 02 09:41:47 crc kubenswrapper[4771]: I1002 09:41:47.914700 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2r4wz" podUID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" containerName="registry-server" containerID="cri-o://031ae6d60b824b867644fb346fda6f99ac137a96c5e80252d3c5fa69de0db797" gracePeriod=2 Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.843994 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.920921 4771 generic.go:334] "Generic (PLEG): container finished" podID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" containerID="031ae6d60b824b867644fb346fda6f99ac137a96c5e80252d3c5fa69de0db797" exitCode=0 Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.920971 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2r4wz" Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.920992 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2r4wz" event={"ID":"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d","Type":"ContainerDied","Data":"031ae6d60b824b867644fb346fda6f99ac137a96c5e80252d3c5fa69de0db797"} Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.921051 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2r4wz" event={"ID":"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d","Type":"ContainerDied","Data":"2c9947b9ff03254723557a9c0838a4e62747ae2168554de9c4aa45ed17f9573b"} Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.921078 4771 scope.go:117] "RemoveContainer" containerID="031ae6d60b824b867644fb346fda6f99ac137a96c5e80252d3c5fa69de0db797" Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.938024 4771 scope.go:117] "RemoveContainer" containerID="9d24821999a730acde162c699127d7d5ae473270e9f51cd8afc12e1ce7036272" Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.953583 4771 scope.go:117] "RemoveContainer" containerID="c624cb5c98fe498d6e7d9aa31abe6b73d3548fe0c047100744a5b67d920d655b" Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.969674 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-catalog-content\") pod \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\" (UID: \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\") " Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.969874 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-utilities\") pod \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\" (UID: \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\") " Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.970423 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlr7b\" (UniqueName: \"kubernetes.io/projected/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-kube-api-access-zlr7b\") pod \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\" (UID: \"ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d\") " Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.970626 4771 scope.go:117] "RemoveContainer" containerID="031ae6d60b824b867644fb346fda6f99ac137a96c5e80252d3c5fa69de0db797" Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.971033 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-utilities" (OuterVolumeSpecName: "utilities") pod "ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" (UID: "ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:48 crc kubenswrapper[4771]: E1002 09:41:48.971117 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"031ae6d60b824b867644fb346fda6f99ac137a96c5e80252d3c5fa69de0db797\": container with ID starting with 031ae6d60b824b867644fb346fda6f99ac137a96c5e80252d3c5fa69de0db797 not found: ID does not exist" containerID="031ae6d60b824b867644fb346fda6f99ac137a96c5e80252d3c5fa69de0db797" Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.971161 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"031ae6d60b824b867644fb346fda6f99ac137a96c5e80252d3c5fa69de0db797"} err="failed to get container status \"031ae6d60b824b867644fb346fda6f99ac137a96c5e80252d3c5fa69de0db797\": rpc error: code = NotFound desc = could not find container \"031ae6d60b824b867644fb346fda6f99ac137a96c5e80252d3c5fa69de0db797\": container with ID starting with 031ae6d60b824b867644fb346fda6f99ac137a96c5e80252d3c5fa69de0db797 not found: ID does not exist" Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.971183 4771 scope.go:117] "RemoveContainer" containerID="9d24821999a730acde162c699127d7d5ae473270e9f51cd8afc12e1ce7036272" Oct 02 09:41:48 crc kubenswrapper[4771]: E1002 09:41:48.971410 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d24821999a730acde162c699127d7d5ae473270e9f51cd8afc12e1ce7036272\": container with ID starting with 9d24821999a730acde162c699127d7d5ae473270e9f51cd8afc12e1ce7036272 not found: ID does not exist" containerID="9d24821999a730acde162c699127d7d5ae473270e9f51cd8afc12e1ce7036272" Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.971442 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d24821999a730acde162c699127d7d5ae473270e9f51cd8afc12e1ce7036272"} err="failed to get container status \"9d24821999a730acde162c699127d7d5ae473270e9f51cd8afc12e1ce7036272\": rpc error: code = NotFound desc = could not find container \"9d24821999a730acde162c699127d7d5ae473270e9f51cd8afc12e1ce7036272\": container with ID starting with 9d24821999a730acde162c699127d7d5ae473270e9f51cd8afc12e1ce7036272 not found: ID does not exist" Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.971467 4771 scope.go:117] "RemoveContainer" containerID="c624cb5c98fe498d6e7d9aa31abe6b73d3548fe0c047100744a5b67d920d655b" Oct 02 09:41:48 crc kubenswrapper[4771]: E1002 09:41:48.971725 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c624cb5c98fe498d6e7d9aa31abe6b73d3548fe0c047100744a5b67d920d655b\": container with ID starting with c624cb5c98fe498d6e7d9aa31abe6b73d3548fe0c047100744a5b67d920d655b not found: ID does not exist" containerID="c624cb5c98fe498d6e7d9aa31abe6b73d3548fe0c047100744a5b67d920d655b" Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.971751 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c624cb5c98fe498d6e7d9aa31abe6b73d3548fe0c047100744a5b67d920d655b"} err="failed to get container status \"c624cb5c98fe498d6e7d9aa31abe6b73d3548fe0c047100744a5b67d920d655b\": rpc error: code = NotFound desc = could not find container \"c624cb5c98fe498d6e7d9aa31abe6b73d3548fe0c047100744a5b67d920d655b\": container with ID starting with c624cb5c98fe498d6e7d9aa31abe6b73d3548fe0c047100744a5b67d920d655b not found: ID does not exist" Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.976672 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-kube-api-access-zlr7b" (OuterVolumeSpecName: "kube-api-access-zlr7b") pod "ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" (UID: "ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d"). InnerVolumeSpecName "kube-api-access-zlr7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:41:48 crc kubenswrapper[4771]: I1002 09:41:48.984047 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" (UID: "ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:49 crc kubenswrapper[4771]: I1002 09:41:49.072024 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:49 crc kubenswrapper[4771]: I1002 09:41:49.072064 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:49 crc kubenswrapper[4771]: I1002 09:41:49.072075 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlr7b\" (UniqueName: \"kubernetes.io/projected/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d-kube-api-access-zlr7b\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:49 crc kubenswrapper[4771]: I1002 09:41:49.250107 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2r4wz"] Oct 02 09:41:49 crc kubenswrapper[4771]: I1002 09:41:49.254568 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2r4wz"] Oct 02 09:41:49 crc kubenswrapper[4771]: I1002 09:41:49.687183 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" path="/var/lib/kubelet/pods/ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d/volumes" Oct 02 09:42:11 crc kubenswrapper[4771]: I1002 09:42:11.955409 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" podUID="e3d49353-2942-437a-bfbb-2da3e5df246d" containerName="oauth-openshift" containerID="cri-o://07dfd1351c030d8adeae28d10382941b51ee980774e191ae260cc34793f5f7d7" gracePeriod=15 Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.271589 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.300754 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-758c4c8f95-zdr4g"] Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301394 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af193d2e-b44e-4519-84c3-cd40eb9eae02" containerName="extract-content" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301414 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="af193d2e-b44e-4519-84c3-cd40eb9eae02" containerName="extract-content" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301426 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301433 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301443 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5107994-4460-4ac0-b975-6ba0fc73a542" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301450 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5107994-4460-4ac0-b975-6ba0fc73a542" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301477 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def237c1-6f8f-40a2-9533-8854ba5079d5" containerName="pruner" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301484 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="def237c1-6f8f-40a2-9533-8854ba5079d5" containerName="pruner" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301491 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a072c8af-e1fb-4af2-98ba-e023152b4cfd" containerName="pruner" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301497 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a072c8af-e1fb-4af2-98ba-e023152b4cfd" containerName="pruner" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301507 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" containerName="extract-utilities" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301513 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" containerName="extract-utilities" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301523 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5107994-4460-4ac0-b975-6ba0fc73a542" containerName="extract-utilities" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301529 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5107994-4460-4ac0-b975-6ba0fc73a542" containerName="extract-utilities" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301559 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af193d2e-b44e-4519-84c3-cd40eb9eae02" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301565 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="af193d2e-b44e-4519-84c3-cd40eb9eae02" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301572 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301578 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301586 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3d49353-2942-437a-bfbb-2da3e5df246d" containerName="oauth-openshift" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301592 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3d49353-2942-437a-bfbb-2da3e5df246d" containerName="oauth-openshift" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301603 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5107994-4460-4ac0-b975-6ba0fc73a542" containerName="extract-content" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301608 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5107994-4460-4ac0-b975-6ba0fc73a542" containerName="extract-content" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301616 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" containerName="extract-utilities" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301640 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" containerName="extract-utilities" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301648 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" containerName="extract-content" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301653 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" containerName="extract-content" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301661 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af193d2e-b44e-4519-84c3-cd40eb9eae02" containerName="extract-utilities" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301667 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="af193d2e-b44e-4519-84c3-cd40eb9eae02" containerName="extract-utilities" Oct 02 09:42:12 crc kubenswrapper[4771]: E1002 09:42:12.301674 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" containerName="extract-content" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301679 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" containerName="extract-content" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301806 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="def237c1-6f8f-40a2-9533-8854ba5079d5" containerName="pruner" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301818 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="d01d796c-0c02-4c8d-b423-e32fc4b187d6" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301824 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="af193d2e-b44e-4519-84c3-cd40eb9eae02" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301834 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a072c8af-e1fb-4af2-98ba-e023152b4cfd" containerName="pruner" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301843 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff77f4a7-a46a-4ff0-9cfb-0aa5cbf79c3d" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301848 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3d49353-2942-437a-bfbb-2da3e5df246d" containerName="oauth-openshift" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.301900 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5107994-4460-4ac0-b975-6ba0fc73a542" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.302315 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.332924 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-758c4c8f95-zdr4g"] Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.385158 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tntj5\" (UniqueName: \"kubernetes.io/projected/e3d49353-2942-437a-bfbb-2da3e5df246d-kube-api-access-tntj5\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.385257 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-cliconfig\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386081 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386130 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-service-ca\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386177 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-idp-0-file-data\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386197 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-ocp-branding-template\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386256 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-serving-cert\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386275 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-session\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386299 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-audit-policies\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386316 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-login\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386351 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-provider-selection\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386373 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-trusted-ca-bundle\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386390 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-router-certs\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386406 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e3d49353-2942-437a-bfbb-2da3e5df246d-audit-dir\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386431 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-error\") pod \"e3d49353-2942-437a-bfbb-2da3e5df246d\" (UID: \"e3d49353-2942-437a-bfbb-2da3e5df246d\") " Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386584 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-router-certs\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386623 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386642 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjkw8\" (UniqueName: \"kubernetes.io/projected/2f95bb51-4623-4d10-8279-358371402464-kube-api-access-cjkw8\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386681 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-session\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386708 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-user-template-error\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386724 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386744 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2f95bb51-4623-4d10-8279-358371402464-audit-dir\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386766 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-serving-cert\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386809 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-cliconfig\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386841 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-service-ca\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386874 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2f95bb51-4623-4d10-8279-358371402464-audit-policies\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386893 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-user-template-login\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386919 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.386937 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.387766 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.387999 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.387954 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3d49353-2942-437a-bfbb-2da3e5df246d-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.388364 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.392672 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.392916 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.392848 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3d49353-2942-437a-bfbb-2da3e5df246d-kube-api-access-tntj5" (OuterVolumeSpecName: "kube-api-access-tntj5") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "kube-api-access-tntj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.393094 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.393302 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.393464 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.393692 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.396618 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.397124 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "e3d49353-2942-437a-bfbb-2da3e5df246d" (UID: "e3d49353-2942-437a-bfbb-2da3e5df246d"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.487946 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2f95bb51-4623-4d10-8279-358371402464-audit-policies\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488000 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-user-template-login\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488032 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488053 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488075 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-router-certs\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488094 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488110 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjkw8\" (UniqueName: \"kubernetes.io/projected/2f95bb51-4623-4d10-8279-358371402464-kube-api-access-cjkw8\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488132 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-session\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488206 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-user-template-error\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488225 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488243 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2f95bb51-4623-4d10-8279-358371402464-audit-dir\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488262 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-serving-cert\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488286 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-cliconfig\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488306 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-service-ca\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488351 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488363 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488373 4771 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488381 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488393 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488402 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488412 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488422 4771 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e3d49353-2942-437a-bfbb-2da3e5df246d-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488431 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488440 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tntj5\" (UniqueName: \"kubernetes.io/projected/e3d49353-2942-437a-bfbb-2da3e5df246d-kube-api-access-tntj5\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488450 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488458 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488466 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.488475 4771 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e3d49353-2942-437a-bfbb-2da3e5df246d-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.489017 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2f95bb51-4623-4d10-8279-358371402464-audit-policies\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.489391 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.489515 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-service-ca\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.490242 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-cliconfig\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.490289 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2f95bb51-4623-4d10-8279-358371402464-audit-dir\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.492317 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.492396 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-user-template-error\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.492482 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-session\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.492516 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-serving-cert\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.492748 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-user-template-login\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.493109 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.493259 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-router-certs\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.496983 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2f95bb51-4623-4d10-8279-358371402464-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.503942 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjkw8\" (UniqueName: \"kubernetes.io/projected/2f95bb51-4623-4d10-8279-358371402464-kube-api-access-cjkw8\") pod \"oauth-openshift-758c4c8f95-zdr4g\" (UID: \"2f95bb51-4623-4d10-8279-358371402464\") " pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.623929 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:12 crc kubenswrapper[4771]: I1002 09:42:12.781752 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-758c4c8f95-zdr4g"] Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.041746 4771 generic.go:334] "Generic (PLEG): container finished" podID="e3d49353-2942-437a-bfbb-2da3e5df246d" containerID="07dfd1351c030d8adeae28d10382941b51ee980774e191ae260cc34793f5f7d7" exitCode=0 Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.041958 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" event={"ID":"e3d49353-2942-437a-bfbb-2da3e5df246d","Type":"ContainerDied","Data":"07dfd1351c030d8adeae28d10382941b51ee980774e191ae260cc34793f5f7d7"} Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.042238 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" event={"ID":"e3d49353-2942-437a-bfbb-2da3e5df246d","Type":"ContainerDied","Data":"0057b189ebe55c1ddfe4c29b2de981b955ab0afac5a894c5fabbe757cecd9762"} Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.042014 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-gdzf9" Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.042265 4771 scope.go:117] "RemoveContainer" containerID="07dfd1351c030d8adeae28d10382941b51ee980774e191ae260cc34793f5f7d7" Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.045181 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" event={"ID":"2f95bb51-4623-4d10-8279-358371402464","Type":"ContainerStarted","Data":"4fff862435f8c62232144efd775849021a5899b25d15338ae1a0a345b1d18190"} Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.045230 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" event={"ID":"2f95bb51-4623-4d10-8279-358371402464","Type":"ContainerStarted","Data":"3006713b08bff57b9a135a0ddf50c1e7f410e06013b13de893f64010285e3649"} Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.045671 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.061049 4771 scope.go:117] "RemoveContainer" containerID="07dfd1351c030d8adeae28d10382941b51ee980774e191ae260cc34793f5f7d7" Oct 02 09:42:13 crc kubenswrapper[4771]: E1002 09:42:13.061680 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07dfd1351c030d8adeae28d10382941b51ee980774e191ae260cc34793f5f7d7\": container with ID starting with 07dfd1351c030d8adeae28d10382941b51ee980774e191ae260cc34793f5f7d7 not found: ID does not exist" containerID="07dfd1351c030d8adeae28d10382941b51ee980774e191ae260cc34793f5f7d7" Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.061711 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07dfd1351c030d8adeae28d10382941b51ee980774e191ae260cc34793f5f7d7"} err="failed to get container status \"07dfd1351c030d8adeae28d10382941b51ee980774e191ae260cc34793f5f7d7\": rpc error: code = NotFound desc = could not find container \"07dfd1351c030d8adeae28d10382941b51ee980774e191ae260cc34793f5f7d7\": container with ID starting with 07dfd1351c030d8adeae28d10382941b51ee980774e191ae260cc34793f5f7d7 not found: ID does not exist" Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.074694 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" podStartSLOduration=27.074672981 podStartE2EDuration="27.074672981s" podCreationTimestamp="2025-10-02 09:41:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:42:13.062464471 +0000 UTC m=+320.710149558" watchObservedRunningTime="2025-10-02 09:42:13.074672981 +0000 UTC m=+320.722358048" Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.082681 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-gdzf9"] Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.086003 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-gdzf9"] Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.352254 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-758c4c8f95-zdr4g" Oct 02 09:42:13 crc kubenswrapper[4771]: I1002 09:42:13.688419 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3d49353-2942-437a-bfbb-2da3e5df246d" path="/var/lib/kubelet/pods/e3d49353-2942-437a-bfbb-2da3e5df246d/volumes" Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.920508 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l5tmt"] Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.921407 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l5tmt" podUID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" containerName="registry-server" containerID="cri-o://9379c2b224c899e1c699864e501137b5c1b464c50a2ec55f83668abab770b06d" gracePeriod=30 Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.941522 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lmpwh"] Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.942389 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lmpwh" podUID="d198e457-892f-40a9-9c39-46e4e67b79a8" containerName="registry-server" containerID="cri-o://48a320f73eb7c7c10577a2f7879c39ac20182fc4ac06befaa2c8436de189c3f4" gracePeriod=30 Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.953210 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cx2j5"] Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.953460 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" podUID="5c65c130-1f41-42f8-be1a-9c8f17bf16dc" containerName="marketplace-operator" containerID="cri-o://ea79706ce5283488be0840742616fa2605002ad5a3acb7809e6bddbe76075725" gracePeriod=30 Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.958598 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qd4wr"] Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.958863 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qd4wr" podUID="820cd711-01de-4cb9-8dac-9cf33c165deb" containerName="registry-server" containerID="cri-o://c16893f4f3f4fd1d2d368e83e1122bc898940a6c79df8ffbc0a63a8a0ed443e6" gracePeriod=30 Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.965271 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d75tj"] Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.965494 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d75tj" podUID="2e959f55-3b26-412b-9128-b9c26fd835b5" containerName="registry-server" containerID="cri-o://1d329a63150ec9fd6a8518cfbe7406ec5b2f866c918675cda2e281c58533db94" gracePeriod=30 Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.976043 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v7w6j"] Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.976934 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.980706 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v7w6j"] Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.997723 4771 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cx2j5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Oct 02 09:42:26 crc kubenswrapper[4771]: I1002 09:42:26.997773 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" podUID="5c65c130-1f41-42f8-be1a-9c8f17bf16dc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.077701 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v7w6j\" (UID: \"36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.078059 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v7w6j\" (UID: \"36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.078096 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfnf8\" (UniqueName: \"kubernetes.io/projected/36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78-kube-api-access-tfnf8\") pod \"marketplace-operator-79b997595-v7w6j\" (UID: \"36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.132077 4771 generic.go:334] "Generic (PLEG): container finished" podID="820cd711-01de-4cb9-8dac-9cf33c165deb" containerID="c16893f4f3f4fd1d2d368e83e1122bc898940a6c79df8ffbc0a63a8a0ed443e6" exitCode=0 Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.132173 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qd4wr" event={"ID":"820cd711-01de-4cb9-8dac-9cf33c165deb","Type":"ContainerDied","Data":"c16893f4f3f4fd1d2d368e83e1122bc898940a6c79df8ffbc0a63a8a0ed443e6"} Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.134183 4771 generic.go:334] "Generic (PLEG): container finished" podID="d198e457-892f-40a9-9c39-46e4e67b79a8" containerID="48a320f73eb7c7c10577a2f7879c39ac20182fc4ac06befaa2c8436de189c3f4" exitCode=0 Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.134258 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lmpwh" event={"ID":"d198e457-892f-40a9-9c39-46e4e67b79a8","Type":"ContainerDied","Data":"48a320f73eb7c7c10577a2f7879c39ac20182fc4ac06befaa2c8436de189c3f4"} Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.135945 4771 generic.go:334] "Generic (PLEG): container finished" podID="2e959f55-3b26-412b-9128-b9c26fd835b5" containerID="1d329a63150ec9fd6a8518cfbe7406ec5b2f866c918675cda2e281c58533db94" exitCode=0 Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.135983 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d75tj" event={"ID":"2e959f55-3b26-412b-9128-b9c26fd835b5","Type":"ContainerDied","Data":"1d329a63150ec9fd6a8518cfbe7406ec5b2f866c918675cda2e281c58533db94"} Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.139705 4771 generic.go:334] "Generic (PLEG): container finished" podID="5c65c130-1f41-42f8-be1a-9c8f17bf16dc" containerID="ea79706ce5283488be0840742616fa2605002ad5a3acb7809e6bddbe76075725" exitCode=0 Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.139791 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" event={"ID":"5c65c130-1f41-42f8-be1a-9c8f17bf16dc","Type":"ContainerDied","Data":"ea79706ce5283488be0840742616fa2605002ad5a3acb7809e6bddbe76075725"} Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.142609 4771 generic.go:334] "Generic (PLEG): container finished" podID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" containerID="9379c2b224c899e1c699864e501137b5c1b464c50a2ec55f83668abab770b06d" exitCode=0 Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.142661 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l5tmt" event={"ID":"8fcbdb41-18a3-4977-aa8b-28f5af63576f","Type":"ContainerDied","Data":"9379c2b224c899e1c699864e501137b5c1b464c50a2ec55f83668abab770b06d"} Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.180186 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v7w6j\" (UID: \"36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.180282 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v7w6j\" (UID: \"36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.180334 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfnf8\" (UniqueName: \"kubernetes.io/projected/36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78-kube-api-access-tfnf8\") pod \"marketplace-operator-79b997595-v7w6j\" (UID: \"36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.182544 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v7w6j\" (UID: \"36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.193986 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v7w6j\" (UID: \"36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.203154 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfnf8\" (UniqueName: \"kubernetes.io/projected/36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78-kube-api-access-tfnf8\") pod \"marketplace-operator-79b997595-v7w6j\" (UID: \"36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.355030 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.364271 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.372852 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.399165 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.404486 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.438339 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.484434 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bm85t\" (UniqueName: \"kubernetes.io/projected/d198e457-892f-40a9-9c39-46e4e67b79a8-kube-api-access-bm85t\") pod \"d198e457-892f-40a9-9c39-46e4e67b79a8\" (UID: \"d198e457-892f-40a9-9c39-46e4e67b79a8\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.484793 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fcbdb41-18a3-4977-aa8b-28f5af63576f-utilities\") pod \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\" (UID: \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.484823 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/820cd711-01de-4cb9-8dac-9cf33c165deb-utilities\") pod \"820cd711-01de-4cb9-8dac-9cf33c165deb\" (UID: \"820cd711-01de-4cb9-8dac-9cf33c165deb\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.484854 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d198e457-892f-40a9-9c39-46e4e67b79a8-utilities\") pod \"d198e457-892f-40a9-9c39-46e4e67b79a8\" (UID: \"d198e457-892f-40a9-9c39-46e4e67b79a8\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.484871 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/820cd711-01de-4cb9-8dac-9cf33c165deb-catalog-content\") pod \"820cd711-01de-4cb9-8dac-9cf33c165deb\" (UID: \"820cd711-01de-4cb9-8dac-9cf33c165deb\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.484892 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fcbdb41-18a3-4977-aa8b-28f5af63576f-catalog-content\") pod \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\" (UID: \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.484926 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp5rh\" (UniqueName: \"kubernetes.io/projected/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-kube-api-access-bp5rh\") pod \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\" (UID: \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.484952 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bpzm\" (UniqueName: \"kubernetes.io/projected/8fcbdb41-18a3-4977-aa8b-28f5af63576f-kube-api-access-8bpzm\") pod \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\" (UID: \"8fcbdb41-18a3-4977-aa8b-28f5af63576f\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.484975 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-marketplace-trusted-ca\") pod \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\" (UID: \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.484998 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-marketplace-operator-metrics\") pod \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\" (UID: \"5c65c130-1f41-42f8-be1a-9c8f17bf16dc\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.485020 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvtm7\" (UniqueName: \"kubernetes.io/projected/820cd711-01de-4cb9-8dac-9cf33c165deb-kube-api-access-qvtm7\") pod \"820cd711-01de-4cb9-8dac-9cf33c165deb\" (UID: \"820cd711-01de-4cb9-8dac-9cf33c165deb\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.485036 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d198e457-892f-40a9-9c39-46e4e67b79a8-catalog-content\") pod \"d198e457-892f-40a9-9c39-46e4e67b79a8\" (UID: \"d198e457-892f-40a9-9c39-46e4e67b79a8\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.487794 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d198e457-892f-40a9-9c39-46e4e67b79a8-utilities" (OuterVolumeSpecName: "utilities") pod "d198e457-892f-40a9-9c39-46e4e67b79a8" (UID: "d198e457-892f-40a9-9c39-46e4e67b79a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.488518 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fcbdb41-18a3-4977-aa8b-28f5af63576f-utilities" (OuterVolumeSpecName: "utilities") pod "8fcbdb41-18a3-4977-aa8b-28f5af63576f" (UID: "8fcbdb41-18a3-4977-aa8b-28f5af63576f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.488522 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "5c65c130-1f41-42f8-be1a-9c8f17bf16dc" (UID: "5c65c130-1f41-42f8-be1a-9c8f17bf16dc"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.489816 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fcbdb41-18a3-4977-aa8b-28f5af63576f-kube-api-access-8bpzm" (OuterVolumeSpecName: "kube-api-access-8bpzm") pod "8fcbdb41-18a3-4977-aa8b-28f5af63576f" (UID: "8fcbdb41-18a3-4977-aa8b-28f5af63576f"). InnerVolumeSpecName "kube-api-access-8bpzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.490119 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/820cd711-01de-4cb9-8dac-9cf33c165deb-kube-api-access-qvtm7" (OuterVolumeSpecName: "kube-api-access-qvtm7") pod "820cd711-01de-4cb9-8dac-9cf33c165deb" (UID: "820cd711-01de-4cb9-8dac-9cf33c165deb"). InnerVolumeSpecName "kube-api-access-qvtm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.490614 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/820cd711-01de-4cb9-8dac-9cf33c165deb-utilities" (OuterVolumeSpecName: "utilities") pod "820cd711-01de-4cb9-8dac-9cf33c165deb" (UID: "820cd711-01de-4cb9-8dac-9cf33c165deb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.493591 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "5c65c130-1f41-42f8-be1a-9c8f17bf16dc" (UID: "5c65c130-1f41-42f8-be1a-9c8f17bf16dc"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.497068 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-kube-api-access-bp5rh" (OuterVolumeSpecName: "kube-api-access-bp5rh") pod "5c65c130-1f41-42f8-be1a-9c8f17bf16dc" (UID: "5c65c130-1f41-42f8-be1a-9c8f17bf16dc"). InnerVolumeSpecName "kube-api-access-bp5rh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.504402 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d198e457-892f-40a9-9c39-46e4e67b79a8-kube-api-access-bm85t" (OuterVolumeSpecName: "kube-api-access-bm85t") pod "d198e457-892f-40a9-9c39-46e4e67b79a8" (UID: "d198e457-892f-40a9-9c39-46e4e67b79a8"). InnerVolumeSpecName "kube-api-access-bm85t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.509359 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/820cd711-01de-4cb9-8dac-9cf33c165deb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "820cd711-01de-4cb9-8dac-9cf33c165deb" (UID: "820cd711-01de-4cb9-8dac-9cf33c165deb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.551852 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fcbdb41-18a3-4977-aa8b-28f5af63576f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8fcbdb41-18a3-4977-aa8b-28f5af63576f" (UID: "8fcbdb41-18a3-4977-aa8b-28f5af63576f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.567979 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d198e457-892f-40a9-9c39-46e4e67b79a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d198e457-892f-40a9-9c39-46e4e67b79a8" (UID: "d198e457-892f-40a9-9c39-46e4e67b79a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586077 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qrmx\" (UniqueName: \"kubernetes.io/projected/2e959f55-3b26-412b-9128-b9c26fd835b5-kube-api-access-5qrmx\") pod \"2e959f55-3b26-412b-9128-b9c26fd835b5\" (UID: \"2e959f55-3b26-412b-9128-b9c26fd835b5\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586255 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e959f55-3b26-412b-9128-b9c26fd835b5-catalog-content\") pod \"2e959f55-3b26-412b-9128-b9c26fd835b5\" (UID: \"2e959f55-3b26-412b-9128-b9c26fd835b5\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586385 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e959f55-3b26-412b-9128-b9c26fd835b5-utilities\") pod \"2e959f55-3b26-412b-9128-b9c26fd835b5\" (UID: \"2e959f55-3b26-412b-9128-b9c26fd835b5\") " Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586635 4771 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586652 4771 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586666 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvtm7\" (UniqueName: \"kubernetes.io/projected/820cd711-01de-4cb9-8dac-9cf33c165deb-kube-api-access-qvtm7\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586681 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d198e457-892f-40a9-9c39-46e4e67b79a8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586692 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bm85t\" (UniqueName: \"kubernetes.io/projected/d198e457-892f-40a9-9c39-46e4e67b79a8-kube-api-access-bm85t\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586705 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fcbdb41-18a3-4977-aa8b-28f5af63576f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586715 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/820cd711-01de-4cb9-8dac-9cf33c165deb-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586725 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d198e457-892f-40a9-9c39-46e4e67b79a8-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586737 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/820cd711-01de-4cb9-8dac-9cf33c165deb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586750 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fcbdb41-18a3-4977-aa8b-28f5af63576f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586760 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp5rh\" (UniqueName: \"kubernetes.io/projected/5c65c130-1f41-42f8-be1a-9c8f17bf16dc-kube-api-access-bp5rh\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.586772 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bpzm\" (UniqueName: \"kubernetes.io/projected/8fcbdb41-18a3-4977-aa8b-28f5af63576f-kube-api-access-8bpzm\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.587163 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e959f55-3b26-412b-9128-b9c26fd835b5-utilities" (OuterVolumeSpecName: "utilities") pod "2e959f55-3b26-412b-9128-b9c26fd835b5" (UID: "2e959f55-3b26-412b-9128-b9c26fd835b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.589408 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e959f55-3b26-412b-9128-b9c26fd835b5-kube-api-access-5qrmx" (OuterVolumeSpecName: "kube-api-access-5qrmx") pod "2e959f55-3b26-412b-9128-b9c26fd835b5" (UID: "2e959f55-3b26-412b-9128-b9c26fd835b5"). InnerVolumeSpecName "kube-api-access-5qrmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.670821 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e959f55-3b26-412b-9128-b9c26fd835b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e959f55-3b26-412b-9128-b9c26fd835b5" (UID: "2e959f55-3b26-412b-9128-b9c26fd835b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.688355 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qrmx\" (UniqueName: \"kubernetes.io/projected/2e959f55-3b26-412b-9128-b9c26fd835b5-kube-api-access-5qrmx\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.688381 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e959f55-3b26-412b-9128-b9c26fd835b5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.688391 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e959f55-3b26-412b-9128-b9c26fd835b5-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:27 crc kubenswrapper[4771]: I1002 09:42:27.819432 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v7w6j"] Oct 02 09:42:27 crc kubenswrapper[4771]: W1002 09:42:27.819722 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36a9a4c8_f9ed_4ed0_b46c_5bd1760e7a78.slice/crio-1026e6246599693e4df181abb1421ffa6814be672b33cae11d35040c2484a7c6 WatchSource:0}: Error finding container 1026e6246599693e4df181abb1421ffa6814be672b33cae11d35040c2484a7c6: Status 404 returned error can't find the container with id 1026e6246599693e4df181abb1421ffa6814be672b33cae11d35040c2484a7c6 Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.150118 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lmpwh" event={"ID":"d198e457-892f-40a9-9c39-46e4e67b79a8","Type":"ContainerDied","Data":"140977a6f77b0f00975bca943d1141919673b54b23bf5828b010fde8b71b2655"} Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.150222 4771 scope.go:117] "RemoveContainer" containerID="48a320f73eb7c7c10577a2f7879c39ac20182fc4ac06befaa2c8436de189c3f4" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.150260 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lmpwh" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.154722 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d75tj" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.154804 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d75tj" event={"ID":"2e959f55-3b26-412b-9128-b9c26fd835b5","Type":"ContainerDied","Data":"9dd68d8d0319116e4c19b6d06a615202a61fe63516fafeeb6f2e536156615b0a"} Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.157167 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" event={"ID":"5c65c130-1f41-42f8-be1a-9c8f17bf16dc","Type":"ContainerDied","Data":"03d11622d647c0499a85d8d53e22db1afe6bd35bc6a9a3f94e598b6c44103764"} Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.157249 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cx2j5" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.167853 4771 scope.go:117] "RemoveContainer" containerID="6d36f3be8fed9b32e81e651936aa866379a702a39145fcc5d045753973d07baa" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.168267 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l5tmt" event={"ID":"8fcbdb41-18a3-4977-aa8b-28f5af63576f","Type":"ContainerDied","Data":"56ca009950917e386f52ba1ea39d5f765fd9a67915a66cb072caed9c9024f99b"} Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.168305 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l5tmt" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.174226 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qd4wr" event={"ID":"820cd711-01de-4cb9-8dac-9cf33c165deb","Type":"ContainerDied","Data":"462704a71879f2a7c351c6ea78fac04278f802ec87c1624b690ac685fab4da12"} Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.174366 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qd4wr" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.179103 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" event={"ID":"36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78","Type":"ContainerStarted","Data":"92bd57f9900201c7b6c2002e05ac726f03e5a12fdec281d286ee50692dfecd53"} Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.179170 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" event={"ID":"36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78","Type":"ContainerStarted","Data":"1026e6246599693e4df181abb1421ffa6814be672b33cae11d35040c2484a7c6"} Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.181395 4771 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-v7w6j container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.181455 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" podUID="36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.183714 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.190835 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lmpwh"] Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.193533 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lmpwh"] Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.195591 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d75tj"] Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.198229 4771 scope.go:117] "RemoveContainer" containerID="3dcd486e2cff2498e972a9c310d3089828f49f0737d6deeaff7481bbf979c70a" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.200275 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d75tj"] Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.213419 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cx2j5"] Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.218086 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cx2j5"] Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.220404 4771 scope.go:117] "RemoveContainer" containerID="1d329a63150ec9fd6a8518cfbe7406ec5b2f866c918675cda2e281c58533db94" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.224456 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qd4wr"] Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.229892 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qd4wr"] Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.244085 4771 scope.go:117] "RemoveContainer" containerID="ea400f0bf9a2031cc4cb0b4168fe1dc30adeb92ef43f4418b4dba2dceaa205d0" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.256457 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" podStartSLOduration=2.256434899 podStartE2EDuration="2.256434899s" podCreationTimestamp="2025-10-02 09:42:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:42:28.244452558 +0000 UTC m=+335.892137625" watchObservedRunningTime="2025-10-02 09:42:28.256434899 +0000 UTC m=+335.904119956" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.256627 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l5tmt"] Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.264766 4771 scope.go:117] "RemoveContainer" containerID="a3f4166383fe85e891dd2b644db2dc3b5a3bd16f720bb500a14f85d454837048" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.265443 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l5tmt"] Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.286031 4771 scope.go:117] "RemoveContainer" containerID="ea79706ce5283488be0840742616fa2605002ad5a3acb7809e6bddbe76075725" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.306613 4771 scope.go:117] "RemoveContainer" containerID="9379c2b224c899e1c699864e501137b5c1b464c50a2ec55f83668abab770b06d" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.323500 4771 scope.go:117] "RemoveContainer" containerID="0f8692d0baddc9926fc14493c2eaa9a9675b5c943cd46bdf4a17dcdf2be5e31d" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.350813 4771 scope.go:117] "RemoveContainer" containerID="84897f45d59677d6d7babd051592e7e97f5a30e7b3c51ce469c748e6358b6638" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.367381 4771 scope.go:117] "RemoveContainer" containerID="c16893f4f3f4fd1d2d368e83e1122bc898940a6c79df8ffbc0a63a8a0ed443e6" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.380230 4771 scope.go:117] "RemoveContainer" containerID="e6389a097fc31f5a74c3d90d8d1ee7a1af07f6571bc0ea3c0090a1fa9ca2ddd1" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.394611 4771 scope.go:117] "RemoveContainer" containerID="7bbe45fcd3465f80cd94debd6a11af5b18fdb65ab3163cfcd9fc3f760be5f160" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941339 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-grhhm"] Oct 02 09:42:28 crc kubenswrapper[4771]: E1002 09:42:28.941543 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d198e457-892f-40a9-9c39-46e4e67b79a8" containerName="registry-server" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941556 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d198e457-892f-40a9-9c39-46e4e67b79a8" containerName="registry-server" Oct 02 09:42:28 crc kubenswrapper[4771]: E1002 09:42:28.941570 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" containerName="extract-content" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941576 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" containerName="extract-content" Oct 02 09:42:28 crc kubenswrapper[4771]: E1002 09:42:28.941587 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="820cd711-01de-4cb9-8dac-9cf33c165deb" containerName="extract-utilities" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941594 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="820cd711-01de-4cb9-8dac-9cf33c165deb" containerName="extract-utilities" Oct 02 09:42:28 crc kubenswrapper[4771]: E1002 09:42:28.941606 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c65c130-1f41-42f8-be1a-9c8f17bf16dc" containerName="marketplace-operator" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941614 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c65c130-1f41-42f8-be1a-9c8f17bf16dc" containerName="marketplace-operator" Oct 02 09:42:28 crc kubenswrapper[4771]: E1002 09:42:28.941623 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" containerName="extract-utilities" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941629 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" containerName="extract-utilities" Oct 02 09:42:28 crc kubenswrapper[4771]: E1002 09:42:28.941636 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e959f55-3b26-412b-9128-b9c26fd835b5" containerName="extract-utilities" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941641 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e959f55-3b26-412b-9128-b9c26fd835b5" containerName="extract-utilities" Oct 02 09:42:28 crc kubenswrapper[4771]: E1002 09:42:28.941648 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="820cd711-01de-4cb9-8dac-9cf33c165deb" containerName="registry-server" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941653 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="820cd711-01de-4cb9-8dac-9cf33c165deb" containerName="registry-server" Oct 02 09:42:28 crc kubenswrapper[4771]: E1002 09:42:28.941664 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d198e457-892f-40a9-9c39-46e4e67b79a8" containerName="extract-content" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941669 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d198e457-892f-40a9-9c39-46e4e67b79a8" containerName="extract-content" Oct 02 09:42:28 crc kubenswrapper[4771]: E1002 09:42:28.941676 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e959f55-3b26-412b-9128-b9c26fd835b5" containerName="registry-server" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941682 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e959f55-3b26-412b-9128-b9c26fd835b5" containerName="registry-server" Oct 02 09:42:28 crc kubenswrapper[4771]: E1002 09:42:28.941691 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="820cd711-01de-4cb9-8dac-9cf33c165deb" containerName="extract-content" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941696 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="820cd711-01de-4cb9-8dac-9cf33c165deb" containerName="extract-content" Oct 02 09:42:28 crc kubenswrapper[4771]: E1002 09:42:28.941704 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e959f55-3b26-412b-9128-b9c26fd835b5" containerName="extract-content" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941710 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e959f55-3b26-412b-9128-b9c26fd835b5" containerName="extract-content" Oct 02 09:42:28 crc kubenswrapper[4771]: E1002 09:42:28.941718 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d198e457-892f-40a9-9c39-46e4e67b79a8" containerName="extract-utilities" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941725 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d198e457-892f-40a9-9c39-46e4e67b79a8" containerName="extract-utilities" Oct 02 09:42:28 crc kubenswrapper[4771]: E1002 09:42:28.941735 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" containerName="registry-server" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941740 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" containerName="registry-server" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941818 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e959f55-3b26-412b-9128-b9c26fd835b5" containerName="registry-server" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941829 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="d198e457-892f-40a9-9c39-46e4e67b79a8" containerName="registry-server" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941837 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="820cd711-01de-4cb9-8dac-9cf33c165deb" containerName="registry-server" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941848 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" containerName="registry-server" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.941856 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c65c130-1f41-42f8-be1a-9c8f17bf16dc" containerName="marketplace-operator" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.942754 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.944528 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 09:42:28 crc kubenswrapper[4771]: I1002 09:42:28.949988 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-grhhm"] Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.014913 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/916b5469-2407-426e-ac06-157981e07018-utilities\") pod \"community-operators-grhhm\" (UID: \"916b5469-2407-426e-ac06-157981e07018\") " pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.014965 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tfs5\" (UniqueName: \"kubernetes.io/projected/916b5469-2407-426e-ac06-157981e07018-kube-api-access-8tfs5\") pod \"community-operators-grhhm\" (UID: \"916b5469-2407-426e-ac06-157981e07018\") " pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.014988 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/916b5469-2407-426e-ac06-157981e07018-catalog-content\") pod \"community-operators-grhhm\" (UID: \"916b5469-2407-426e-ac06-157981e07018\") " pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.116372 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/916b5469-2407-426e-ac06-157981e07018-utilities\") pod \"community-operators-grhhm\" (UID: \"916b5469-2407-426e-ac06-157981e07018\") " pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.116459 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tfs5\" (UniqueName: \"kubernetes.io/projected/916b5469-2407-426e-ac06-157981e07018-kube-api-access-8tfs5\") pod \"community-operators-grhhm\" (UID: \"916b5469-2407-426e-ac06-157981e07018\") " pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.116478 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/916b5469-2407-426e-ac06-157981e07018-catalog-content\") pod \"community-operators-grhhm\" (UID: \"916b5469-2407-426e-ac06-157981e07018\") " pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.116888 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/916b5469-2407-426e-ac06-157981e07018-catalog-content\") pod \"community-operators-grhhm\" (UID: \"916b5469-2407-426e-ac06-157981e07018\") " pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.116887 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/916b5469-2407-426e-ac06-157981e07018-utilities\") pod \"community-operators-grhhm\" (UID: \"916b5469-2407-426e-ac06-157981e07018\") " pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.136038 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tfs5\" (UniqueName: \"kubernetes.io/projected/916b5469-2407-426e-ac06-157981e07018-kube-api-access-8tfs5\") pod \"community-operators-grhhm\" (UID: \"916b5469-2407-426e-ac06-157981e07018\") " pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.189495 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-v7w6j" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.269176 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.432925 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-grhhm"] Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.540599 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w8d6f"] Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.542989 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.545461 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.557379 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w8d6f"] Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.623061 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e228deeb-ac11-4ca5-8e08-5395e28bfb07-catalog-content\") pod \"redhat-marketplace-w8d6f\" (UID: \"e228deeb-ac11-4ca5-8e08-5395e28bfb07\") " pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.623153 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptw8q\" (UniqueName: \"kubernetes.io/projected/e228deeb-ac11-4ca5-8e08-5395e28bfb07-kube-api-access-ptw8q\") pod \"redhat-marketplace-w8d6f\" (UID: \"e228deeb-ac11-4ca5-8e08-5395e28bfb07\") " pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.623187 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e228deeb-ac11-4ca5-8e08-5395e28bfb07-utilities\") pod \"redhat-marketplace-w8d6f\" (UID: \"e228deeb-ac11-4ca5-8e08-5395e28bfb07\") " pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.689467 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e959f55-3b26-412b-9128-b9c26fd835b5" path="/var/lib/kubelet/pods/2e959f55-3b26-412b-9128-b9c26fd835b5/volumes" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.690323 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c65c130-1f41-42f8-be1a-9c8f17bf16dc" path="/var/lib/kubelet/pods/5c65c130-1f41-42f8-be1a-9c8f17bf16dc/volumes" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.690958 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="820cd711-01de-4cb9-8dac-9cf33c165deb" path="/var/lib/kubelet/pods/820cd711-01de-4cb9-8dac-9cf33c165deb/volumes" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.692320 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fcbdb41-18a3-4977-aa8b-28f5af63576f" path="/var/lib/kubelet/pods/8fcbdb41-18a3-4977-aa8b-28f5af63576f/volumes" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.693069 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d198e457-892f-40a9-9c39-46e4e67b79a8" path="/var/lib/kubelet/pods/d198e457-892f-40a9-9c39-46e4e67b79a8/volumes" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.724722 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptw8q\" (UniqueName: \"kubernetes.io/projected/e228deeb-ac11-4ca5-8e08-5395e28bfb07-kube-api-access-ptw8q\") pod \"redhat-marketplace-w8d6f\" (UID: \"e228deeb-ac11-4ca5-8e08-5395e28bfb07\") " pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.724787 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e228deeb-ac11-4ca5-8e08-5395e28bfb07-utilities\") pod \"redhat-marketplace-w8d6f\" (UID: \"e228deeb-ac11-4ca5-8e08-5395e28bfb07\") " pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.724852 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e228deeb-ac11-4ca5-8e08-5395e28bfb07-catalog-content\") pod \"redhat-marketplace-w8d6f\" (UID: \"e228deeb-ac11-4ca5-8e08-5395e28bfb07\") " pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.725425 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e228deeb-ac11-4ca5-8e08-5395e28bfb07-catalog-content\") pod \"redhat-marketplace-w8d6f\" (UID: \"e228deeb-ac11-4ca5-8e08-5395e28bfb07\") " pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.725612 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e228deeb-ac11-4ca5-8e08-5395e28bfb07-utilities\") pod \"redhat-marketplace-w8d6f\" (UID: \"e228deeb-ac11-4ca5-8e08-5395e28bfb07\") " pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.747521 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptw8q\" (UniqueName: \"kubernetes.io/projected/e228deeb-ac11-4ca5-8e08-5395e28bfb07-kube-api-access-ptw8q\") pod \"redhat-marketplace-w8d6f\" (UID: \"e228deeb-ac11-4ca5-8e08-5395e28bfb07\") " pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:29 crc kubenswrapper[4771]: I1002 09:42:29.859395 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:30 crc kubenswrapper[4771]: I1002 09:42:30.193252 4771 generic.go:334] "Generic (PLEG): container finished" podID="916b5469-2407-426e-ac06-157981e07018" containerID="7d1844110b499d8e17bf5cdb9b06708a5141587a5d352e13f121e70bd2a4f4b8" exitCode=0 Oct 02 09:42:30 crc kubenswrapper[4771]: I1002 09:42:30.193299 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-grhhm" event={"ID":"916b5469-2407-426e-ac06-157981e07018","Type":"ContainerDied","Data":"7d1844110b499d8e17bf5cdb9b06708a5141587a5d352e13f121e70bd2a4f4b8"} Oct 02 09:42:30 crc kubenswrapper[4771]: I1002 09:42:30.193346 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-grhhm" event={"ID":"916b5469-2407-426e-ac06-157981e07018","Type":"ContainerStarted","Data":"19d84ab752b36e880b67a0e4c1ccf07bca38a2be856b15d0d6e808428ba9cc60"} Oct 02 09:42:30 crc kubenswrapper[4771]: I1002 09:42:30.285380 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w8d6f"] Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.200525 4771 generic.go:334] "Generic (PLEG): container finished" podID="e228deeb-ac11-4ca5-8e08-5395e28bfb07" containerID="8e75a4337fc8175303e11570ab94e02eac4fad7843d2842ab7ebab46fa59f36c" exitCode=0 Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.200638 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w8d6f" event={"ID":"e228deeb-ac11-4ca5-8e08-5395e28bfb07","Type":"ContainerDied","Data":"8e75a4337fc8175303e11570ab94e02eac4fad7843d2842ab7ebab46fa59f36c"} Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.200929 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w8d6f" event={"ID":"e228deeb-ac11-4ca5-8e08-5395e28bfb07","Type":"ContainerStarted","Data":"1f3127b52d89287e7a2a41d35796da441dd4c605a91110fe8392927abae80b60"} Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.345168 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9sfxx"] Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.346241 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.348192 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.354116 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9sfxx"] Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.447071 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9be41903-7359-484f-8e0d-3a94e53ade9a-catalog-content\") pod \"redhat-operators-9sfxx\" (UID: \"9be41903-7359-484f-8e0d-3a94e53ade9a\") " pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.447164 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9be41903-7359-484f-8e0d-3a94e53ade9a-utilities\") pod \"redhat-operators-9sfxx\" (UID: \"9be41903-7359-484f-8e0d-3a94e53ade9a\") " pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.447363 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zb9t\" (UniqueName: \"kubernetes.io/projected/9be41903-7359-484f-8e0d-3a94e53ade9a-kube-api-access-7zb9t\") pod \"redhat-operators-9sfxx\" (UID: \"9be41903-7359-484f-8e0d-3a94e53ade9a\") " pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.548979 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zb9t\" (UniqueName: \"kubernetes.io/projected/9be41903-7359-484f-8e0d-3a94e53ade9a-kube-api-access-7zb9t\") pod \"redhat-operators-9sfxx\" (UID: \"9be41903-7359-484f-8e0d-3a94e53ade9a\") " pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.549112 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9be41903-7359-484f-8e0d-3a94e53ade9a-catalog-content\") pod \"redhat-operators-9sfxx\" (UID: \"9be41903-7359-484f-8e0d-3a94e53ade9a\") " pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.549165 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9be41903-7359-484f-8e0d-3a94e53ade9a-utilities\") pod \"redhat-operators-9sfxx\" (UID: \"9be41903-7359-484f-8e0d-3a94e53ade9a\") " pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.549684 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9be41903-7359-484f-8e0d-3a94e53ade9a-catalog-content\") pod \"redhat-operators-9sfxx\" (UID: \"9be41903-7359-484f-8e0d-3a94e53ade9a\") " pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.549758 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9be41903-7359-484f-8e0d-3a94e53ade9a-utilities\") pod \"redhat-operators-9sfxx\" (UID: \"9be41903-7359-484f-8e0d-3a94e53ade9a\") " pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.573563 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zb9t\" (UniqueName: \"kubernetes.io/projected/9be41903-7359-484f-8e0d-3a94e53ade9a-kube-api-access-7zb9t\") pod \"redhat-operators-9sfxx\" (UID: \"9be41903-7359-484f-8e0d-3a94e53ade9a\") " pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.666448 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.848052 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9sfxx"] Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.941673 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-md7x4"] Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.944333 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.947423 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 09:42:31 crc kubenswrapper[4771]: I1002 09:42:31.952958 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-md7x4"] Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.054764 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c89214fa-0d7c-4937-9cf3-6545a47420b9-utilities\") pod \"certified-operators-md7x4\" (UID: \"c89214fa-0d7c-4937-9cf3-6545a47420b9\") " pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.055293 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c89214fa-0d7c-4937-9cf3-6545a47420b9-catalog-content\") pod \"certified-operators-md7x4\" (UID: \"c89214fa-0d7c-4937-9cf3-6545a47420b9\") " pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.055409 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jg8f\" (UniqueName: \"kubernetes.io/projected/c89214fa-0d7c-4937-9cf3-6545a47420b9-kube-api-access-5jg8f\") pod \"certified-operators-md7x4\" (UID: \"c89214fa-0d7c-4937-9cf3-6545a47420b9\") " pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.157483 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jg8f\" (UniqueName: \"kubernetes.io/projected/c89214fa-0d7c-4937-9cf3-6545a47420b9-kube-api-access-5jg8f\") pod \"certified-operators-md7x4\" (UID: \"c89214fa-0d7c-4937-9cf3-6545a47420b9\") " pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.157567 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c89214fa-0d7c-4937-9cf3-6545a47420b9-utilities\") pod \"certified-operators-md7x4\" (UID: \"c89214fa-0d7c-4937-9cf3-6545a47420b9\") " pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.157591 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c89214fa-0d7c-4937-9cf3-6545a47420b9-catalog-content\") pod \"certified-operators-md7x4\" (UID: \"c89214fa-0d7c-4937-9cf3-6545a47420b9\") " pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.158071 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c89214fa-0d7c-4937-9cf3-6545a47420b9-catalog-content\") pod \"certified-operators-md7x4\" (UID: \"c89214fa-0d7c-4937-9cf3-6545a47420b9\") " pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.158696 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c89214fa-0d7c-4937-9cf3-6545a47420b9-utilities\") pod \"certified-operators-md7x4\" (UID: \"c89214fa-0d7c-4937-9cf3-6545a47420b9\") " pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.182932 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jg8f\" (UniqueName: \"kubernetes.io/projected/c89214fa-0d7c-4937-9cf3-6545a47420b9-kube-api-access-5jg8f\") pod \"certified-operators-md7x4\" (UID: \"c89214fa-0d7c-4937-9cf3-6545a47420b9\") " pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.208197 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w8d6f" event={"ID":"e228deeb-ac11-4ca5-8e08-5395e28bfb07","Type":"ContainerStarted","Data":"c133e72ab9851eeb606817f301cdab54daa36af93d4e035cb07b10d23237ee98"} Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.210972 4771 generic.go:334] "Generic (PLEG): container finished" podID="9be41903-7359-484f-8e0d-3a94e53ade9a" containerID="1c55649bf15d896f209e4509de19f30420110e56cc54b8a6c3ef2b3d412cba3e" exitCode=0 Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.211069 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9sfxx" event={"ID":"9be41903-7359-484f-8e0d-3a94e53ade9a","Type":"ContainerDied","Data":"1c55649bf15d896f209e4509de19f30420110e56cc54b8a6c3ef2b3d412cba3e"} Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.211154 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9sfxx" event={"ID":"9be41903-7359-484f-8e0d-3a94e53ade9a","Type":"ContainerStarted","Data":"91dd985ec8925414a03fb84dc94c9650f2ee6338e95d1d96028141f43e88be53"} Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.214812 4771 generic.go:334] "Generic (PLEG): container finished" podID="916b5469-2407-426e-ac06-157981e07018" containerID="1734add2406869174527b5577e147c940b6f5d7f34f2d88ad3ee39cf7251b757" exitCode=0 Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.214855 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-grhhm" event={"ID":"916b5469-2407-426e-ac06-157981e07018","Type":"ContainerDied","Data":"1734add2406869174527b5577e147c940b6f5d7f34f2d88ad3ee39cf7251b757"} Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.271244 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:32 crc kubenswrapper[4771]: I1002 09:42:32.449434 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-md7x4"] Oct 02 09:42:32 crc kubenswrapper[4771]: W1002 09:42:32.457854 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc89214fa_0d7c_4937_9cf3_6545a47420b9.slice/crio-442e615f09d7b78b3d94f62d0a267b4b6b8b56284be520b5c95ee5f1a8de3701 WatchSource:0}: Error finding container 442e615f09d7b78b3d94f62d0a267b4b6b8b56284be520b5c95ee5f1a8de3701: Status 404 returned error can't find the container with id 442e615f09d7b78b3d94f62d0a267b4b6b8b56284be520b5c95ee5f1a8de3701 Oct 02 09:42:33 crc kubenswrapper[4771]: I1002 09:42:33.223648 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-grhhm" event={"ID":"916b5469-2407-426e-ac06-157981e07018","Type":"ContainerStarted","Data":"172b205944ad3e0218c98d7cb1ad1f3cc7486aa31ac303d9a6d10e7943263b27"} Oct 02 09:42:33 crc kubenswrapper[4771]: I1002 09:42:33.225901 4771 generic.go:334] "Generic (PLEG): container finished" podID="c89214fa-0d7c-4937-9cf3-6545a47420b9" containerID="edabd8c265665fd16d88a38c128ff13a2a69e2f3cd99b16257107bc7615dd584" exitCode=0 Oct 02 09:42:33 crc kubenswrapper[4771]: I1002 09:42:33.225942 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md7x4" event={"ID":"c89214fa-0d7c-4937-9cf3-6545a47420b9","Type":"ContainerDied","Data":"edabd8c265665fd16d88a38c128ff13a2a69e2f3cd99b16257107bc7615dd584"} Oct 02 09:42:33 crc kubenswrapper[4771]: I1002 09:42:33.225957 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md7x4" event={"ID":"c89214fa-0d7c-4937-9cf3-6545a47420b9","Type":"ContainerStarted","Data":"442e615f09d7b78b3d94f62d0a267b4b6b8b56284be520b5c95ee5f1a8de3701"} Oct 02 09:42:33 crc kubenswrapper[4771]: I1002 09:42:33.228392 4771 generic.go:334] "Generic (PLEG): container finished" podID="e228deeb-ac11-4ca5-8e08-5395e28bfb07" containerID="c133e72ab9851eeb606817f301cdab54daa36af93d4e035cb07b10d23237ee98" exitCode=0 Oct 02 09:42:33 crc kubenswrapper[4771]: I1002 09:42:33.228461 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w8d6f" event={"ID":"e228deeb-ac11-4ca5-8e08-5395e28bfb07","Type":"ContainerDied","Data":"c133e72ab9851eeb606817f301cdab54daa36af93d4e035cb07b10d23237ee98"} Oct 02 09:42:33 crc kubenswrapper[4771]: I1002 09:42:33.229780 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9sfxx" event={"ID":"9be41903-7359-484f-8e0d-3a94e53ade9a","Type":"ContainerStarted","Data":"8837dbd9622769884a8f7df7cee296c7f6cf027aa0fded405af23129c8cff9e9"} Oct 02 09:42:33 crc kubenswrapper[4771]: I1002 09:42:33.248710 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-grhhm" podStartSLOduration=2.8097265240000002 podStartE2EDuration="5.248695382s" podCreationTimestamp="2025-10-02 09:42:28 +0000 UTC" firstStartedPulling="2025-10-02 09:42:30.19506603 +0000 UTC m=+337.842751097" lastFinishedPulling="2025-10-02 09:42:32.634034888 +0000 UTC m=+340.281719955" observedRunningTime="2025-10-02 09:42:33.246938186 +0000 UTC m=+340.894623263" watchObservedRunningTime="2025-10-02 09:42:33.248695382 +0000 UTC m=+340.896380449" Oct 02 09:42:34 crc kubenswrapper[4771]: I1002 09:42:34.238177 4771 generic.go:334] "Generic (PLEG): container finished" podID="9be41903-7359-484f-8e0d-3a94e53ade9a" containerID="8837dbd9622769884a8f7df7cee296c7f6cf027aa0fded405af23129c8cff9e9" exitCode=0 Oct 02 09:42:34 crc kubenswrapper[4771]: I1002 09:42:34.238739 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9sfxx" event={"ID":"9be41903-7359-484f-8e0d-3a94e53ade9a","Type":"ContainerDied","Data":"8837dbd9622769884a8f7df7cee296c7f6cf027aa0fded405af23129c8cff9e9"} Oct 02 09:42:34 crc kubenswrapper[4771]: I1002 09:42:34.243439 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md7x4" event={"ID":"c89214fa-0d7c-4937-9cf3-6545a47420b9","Type":"ContainerStarted","Data":"a9fd62b4133d260ff539e5807a23c0a2ce37cdfd9d172899a166ec7ae636141a"} Oct 02 09:42:34 crc kubenswrapper[4771]: I1002 09:42:34.249185 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w8d6f" event={"ID":"e228deeb-ac11-4ca5-8e08-5395e28bfb07","Type":"ContainerStarted","Data":"ddeb1b7bd4094a28710a53906065f1a0c48f16689f75f48e952758687fc07dd3"} Oct 02 09:42:34 crc kubenswrapper[4771]: I1002 09:42:34.274609 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w8d6f" podStartSLOduration=2.7946193790000002 podStartE2EDuration="5.27458429s" podCreationTimestamp="2025-10-02 09:42:29 +0000 UTC" firstStartedPulling="2025-10-02 09:42:31.203390383 +0000 UTC m=+338.851075450" lastFinishedPulling="2025-10-02 09:42:33.683355294 +0000 UTC m=+341.331040361" observedRunningTime="2025-10-02 09:42:34.271622973 +0000 UTC m=+341.919308040" watchObservedRunningTime="2025-10-02 09:42:34.27458429 +0000 UTC m=+341.922269357" Oct 02 09:42:35 crc kubenswrapper[4771]: I1002 09:42:35.257728 4771 generic.go:334] "Generic (PLEG): container finished" podID="c89214fa-0d7c-4937-9cf3-6545a47420b9" containerID="a9fd62b4133d260ff539e5807a23c0a2ce37cdfd9d172899a166ec7ae636141a" exitCode=0 Oct 02 09:42:35 crc kubenswrapper[4771]: I1002 09:42:35.257778 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md7x4" event={"ID":"c89214fa-0d7c-4937-9cf3-6545a47420b9","Type":"ContainerDied","Data":"a9fd62b4133d260ff539e5807a23c0a2ce37cdfd9d172899a166ec7ae636141a"} Oct 02 09:42:35 crc kubenswrapper[4771]: I1002 09:42:35.261617 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9sfxx" event={"ID":"9be41903-7359-484f-8e0d-3a94e53ade9a","Type":"ContainerStarted","Data":"da16917890e3ec383550be2207525ce516277cc94048e373a33ae62c70d028ca"} Oct 02 09:42:36 crc kubenswrapper[4771]: I1002 09:42:36.268696 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md7x4" event={"ID":"c89214fa-0d7c-4937-9cf3-6545a47420b9","Type":"ContainerStarted","Data":"ebb7d67bee28d1b4028e1f0796ab23e71d6eac0f26092e600dc3623f8265ea09"} Oct 02 09:42:36 crc kubenswrapper[4771]: I1002 09:42:36.288272 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9sfxx" podStartSLOduration=2.697221354 podStartE2EDuration="5.288254648s" podCreationTimestamp="2025-10-02 09:42:31 +0000 UTC" firstStartedPulling="2025-10-02 09:42:32.213010659 +0000 UTC m=+339.860695726" lastFinishedPulling="2025-10-02 09:42:34.804043953 +0000 UTC m=+342.451729020" observedRunningTime="2025-10-02 09:42:35.293478707 +0000 UTC m=+342.941163774" watchObservedRunningTime="2025-10-02 09:42:36.288254648 +0000 UTC m=+343.935939715" Oct 02 09:42:36 crc kubenswrapper[4771]: I1002 09:42:36.288428 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-md7x4" podStartSLOduration=2.753928247 podStartE2EDuration="5.288423413s" podCreationTimestamp="2025-10-02 09:42:31 +0000 UTC" firstStartedPulling="2025-10-02 09:42:33.227216035 +0000 UTC m=+340.874901102" lastFinishedPulling="2025-10-02 09:42:35.761711201 +0000 UTC m=+343.409396268" observedRunningTime="2025-10-02 09:42:36.286049271 +0000 UTC m=+343.933734338" watchObservedRunningTime="2025-10-02 09:42:36.288423413 +0000 UTC m=+343.936108480" Oct 02 09:42:39 crc kubenswrapper[4771]: I1002 09:42:39.269430 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:39 crc kubenswrapper[4771]: I1002 09:42:39.270068 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:39 crc kubenswrapper[4771]: I1002 09:42:39.315620 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:39 crc kubenswrapper[4771]: I1002 09:42:39.356055 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-grhhm" Oct 02 09:42:39 crc kubenswrapper[4771]: I1002 09:42:39.861371 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:39 crc kubenswrapper[4771]: I1002 09:42:39.862183 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:39 crc kubenswrapper[4771]: I1002 09:42:39.915605 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:40 crc kubenswrapper[4771]: I1002 09:42:40.327298 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w8d6f" Oct 02 09:42:41 crc kubenswrapper[4771]: I1002 09:42:41.667295 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:41 crc kubenswrapper[4771]: I1002 09:42:41.667877 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:41 crc kubenswrapper[4771]: I1002 09:42:41.708775 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:42 crc kubenswrapper[4771]: I1002 09:42:42.146201 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:42:42 crc kubenswrapper[4771]: I1002 09:42:42.146273 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:42:42 crc kubenswrapper[4771]: I1002 09:42:42.272327 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:42 crc kubenswrapper[4771]: I1002 09:42:42.272417 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:42 crc kubenswrapper[4771]: I1002 09:42:42.314616 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:42 crc kubenswrapper[4771]: I1002 09:42:42.357883 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9sfxx" Oct 02 09:42:42 crc kubenswrapper[4771]: I1002 09:42:42.368405 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-md7x4" Oct 02 09:42:56 crc kubenswrapper[4771]: I1002 09:42:56.877798 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962"] Oct 02 09:42:56 crc kubenswrapper[4771]: I1002 09:42:56.880886 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" Oct 02 09:42:56 crc kubenswrapper[4771]: I1002 09:42:56.883729 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Oct 02 09:42:56 crc kubenswrapper[4771]: I1002 09:42:56.883806 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Oct 02 09:42:56 crc kubenswrapper[4771]: I1002 09:42:56.884083 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Oct 02 09:42:56 crc kubenswrapper[4771]: I1002 09:42:56.887707 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-dockercfg-wwt9l" Oct 02 09:42:56 crc kubenswrapper[4771]: I1002 09:42:56.890162 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962"] Oct 02 09:42:56 crc kubenswrapper[4771]: I1002 09:42:56.890953 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Oct 02 09:42:57 crc kubenswrapper[4771]: I1002 09:42:57.005038 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn8c9\" (UniqueName: \"kubernetes.io/projected/e0931220-10df-49f0-b086-e10e39015d32-kube-api-access-rn8c9\") pod \"cluster-monitoring-operator-6d5b84845-8p962\" (UID: \"e0931220-10df-49f0-b086-e10e39015d32\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" Oct 02 09:42:57 crc kubenswrapper[4771]: I1002 09:42:57.005213 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/e0931220-10df-49f0-b086-e10e39015d32-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-8p962\" (UID: \"e0931220-10df-49f0-b086-e10e39015d32\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" Oct 02 09:42:57 crc kubenswrapper[4771]: I1002 09:42:57.005268 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/e0931220-10df-49f0-b086-e10e39015d32-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-8p962\" (UID: \"e0931220-10df-49f0-b086-e10e39015d32\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" Oct 02 09:42:57 crc kubenswrapper[4771]: I1002 09:42:57.106942 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/e0931220-10df-49f0-b086-e10e39015d32-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-8p962\" (UID: \"e0931220-10df-49f0-b086-e10e39015d32\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" Oct 02 09:42:57 crc kubenswrapper[4771]: I1002 09:42:57.107046 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn8c9\" (UniqueName: \"kubernetes.io/projected/e0931220-10df-49f0-b086-e10e39015d32-kube-api-access-rn8c9\") pod \"cluster-monitoring-operator-6d5b84845-8p962\" (UID: \"e0931220-10df-49f0-b086-e10e39015d32\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" Oct 02 09:42:57 crc kubenswrapper[4771]: I1002 09:42:57.107107 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/e0931220-10df-49f0-b086-e10e39015d32-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-8p962\" (UID: \"e0931220-10df-49f0-b086-e10e39015d32\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" Oct 02 09:42:57 crc kubenswrapper[4771]: I1002 09:42:57.108403 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/e0931220-10df-49f0-b086-e10e39015d32-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-8p962\" (UID: \"e0931220-10df-49f0-b086-e10e39015d32\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" Oct 02 09:42:57 crc kubenswrapper[4771]: I1002 09:42:57.115678 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/e0931220-10df-49f0-b086-e10e39015d32-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-8p962\" (UID: \"e0931220-10df-49f0-b086-e10e39015d32\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" Oct 02 09:42:57 crc kubenswrapper[4771]: I1002 09:42:57.124193 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn8c9\" (UniqueName: \"kubernetes.io/projected/e0931220-10df-49f0-b086-e10e39015d32-kube-api-access-rn8c9\") pod \"cluster-monitoring-operator-6d5b84845-8p962\" (UID: \"e0931220-10df-49f0-b086-e10e39015d32\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" Oct 02 09:42:57 crc kubenswrapper[4771]: I1002 09:42:57.204845 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" Oct 02 09:42:57 crc kubenswrapper[4771]: I1002 09:42:57.589755 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962"] Oct 02 09:42:58 crc kubenswrapper[4771]: I1002 09:42:58.391042 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" event={"ID":"e0931220-10df-49f0-b086-e10e39015d32","Type":"ContainerStarted","Data":"3f8a77dfb22be3967e65145b6d1c4a5d5057e2559fa0e4928f108f1026a29be8"} Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.401348 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" event={"ID":"e0931220-10df-49f0-b086-e10e39015d32","Type":"ContainerStarted","Data":"37f40f2c9cdfa8f2ce1d086d37408798a2aea2ba77ae0b4e6f26f294234ffba4"} Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.418371 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-8p962" podStartSLOduration=1.8037113919999999 podStartE2EDuration="3.418355743s" podCreationTimestamp="2025-10-02 09:42:56 +0000 UTC" firstStartedPulling="2025-10-02 09:42:57.605366205 +0000 UTC m=+365.253051272" lastFinishedPulling="2025-10-02 09:42:59.220010556 +0000 UTC m=+366.867695623" observedRunningTime="2025-10-02 09:42:59.416379892 +0000 UTC m=+367.064064959" watchObservedRunningTime="2025-10-02 09:42:59.418355743 +0000 UTC m=+367.066040810" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.741679 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-pq9fh"] Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.742604 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.760918 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-pq9fh"] Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.844483 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-ca-trust-extracted\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.844556 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-bound-sa-token\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.844760 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.844885 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gzhq\" (UniqueName: \"kubernetes.io/projected/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-kube-api-access-8gzhq\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.844949 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-trusted-ca\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.845228 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-installation-pull-secrets\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.845385 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-registry-tls\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.845426 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-registry-certificates\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.902961 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.914436 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-htfz2"] Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.915490 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-htfz2" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.917586 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-dockercfg-ztvmf" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.917906 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.920507 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-htfz2"] Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.946886 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gzhq\" (UniqueName: \"kubernetes.io/projected/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-kube-api-access-8gzhq\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.946939 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-trusted-ca\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.946967 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-installation-pull-secrets\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.946995 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-registry-tls\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.947013 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-registry-certificates\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.947050 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-ca-trust-extracted\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.947065 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-bound-sa-token\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.947927 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-ca-trust-extracted\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.948594 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-trusted-ca\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.948720 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-registry-certificates\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.964186 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-installation-pull-secrets\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.964314 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-registry-tls\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.967315 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-bound-sa-token\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:42:59 crc kubenswrapper[4771]: I1002 09:42:59.968238 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gzhq\" (UniqueName: \"kubernetes.io/projected/93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea-kube-api-access-8gzhq\") pod \"image-registry-66df7c8f76-pq9fh\" (UID: \"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea\") " pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:43:00 crc kubenswrapper[4771]: I1002 09:43:00.048685 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/e302e18b-9d3f-4332-9ef9-37ac29846033-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-htfz2\" (UID: \"e302e18b-9d3f-4332-9ef9-37ac29846033\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-htfz2" Oct 02 09:43:00 crc kubenswrapper[4771]: I1002 09:43:00.059629 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:43:00 crc kubenswrapper[4771]: I1002 09:43:00.150868 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/e302e18b-9d3f-4332-9ef9-37ac29846033-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-htfz2\" (UID: \"e302e18b-9d3f-4332-9ef9-37ac29846033\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-htfz2" Oct 02 09:43:00 crc kubenswrapper[4771]: I1002 09:43:00.160499 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/e302e18b-9d3f-4332-9ef9-37ac29846033-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-htfz2\" (UID: \"e302e18b-9d3f-4332-9ef9-37ac29846033\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-htfz2" Oct 02 09:43:00 crc kubenswrapper[4771]: I1002 09:43:00.238586 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-htfz2" Oct 02 09:43:00 crc kubenswrapper[4771]: I1002 09:43:00.273583 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-pq9fh"] Oct 02 09:43:00 crc kubenswrapper[4771]: W1002 09:43:00.283334 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93b0147c_a6fe_4ce6_ae6b_fd4d2bd3dcea.slice/crio-b9b6bfef55302b2c1ca7bec95c4125ac8cf745027224e7c91809133ca9a5519a WatchSource:0}: Error finding container b9b6bfef55302b2c1ca7bec95c4125ac8cf745027224e7c91809133ca9a5519a: Status 404 returned error can't find the container with id b9b6bfef55302b2c1ca7bec95c4125ac8cf745027224e7c91809133ca9a5519a Oct 02 09:43:00 crc kubenswrapper[4771]: I1002 09:43:00.409536 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" event={"ID":"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea","Type":"ContainerStarted","Data":"b9b6bfef55302b2c1ca7bec95c4125ac8cf745027224e7c91809133ca9a5519a"} Oct 02 09:43:00 crc kubenswrapper[4771]: I1002 09:43:00.443956 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-htfz2"] Oct 02 09:43:00 crc kubenswrapper[4771]: W1002 09:43:00.448315 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode302e18b_9d3f_4332_9ef9_37ac29846033.slice/crio-cb6294c79685b0a0ade9605cad37cad3db6833740963b7161d48e77cbc77c8f4 WatchSource:0}: Error finding container cb6294c79685b0a0ade9605cad37cad3db6833740963b7161d48e77cbc77c8f4: Status 404 returned error can't find the container with id cb6294c79685b0a0ade9605cad37cad3db6833740963b7161d48e77cbc77c8f4 Oct 02 09:43:01 crc kubenswrapper[4771]: I1002 09:43:01.420038 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" event={"ID":"93b0147c-a6fe-4ce6-ae6b-fd4d2bd3dcea","Type":"ContainerStarted","Data":"c13d8b7104689012cac1c4966ff7acee3d423378d2fe1c07a165c2807c3b761c"} Oct 02 09:43:01 crc kubenswrapper[4771]: I1002 09:43:01.420766 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:43:01 crc kubenswrapper[4771]: I1002 09:43:01.421092 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-htfz2" event={"ID":"e302e18b-9d3f-4332-9ef9-37ac29846033","Type":"ContainerStarted","Data":"cb6294c79685b0a0ade9605cad37cad3db6833740963b7161d48e77cbc77c8f4"} Oct 02 09:43:01 crc kubenswrapper[4771]: I1002 09:43:01.447420 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" podStartSLOduration=2.4473839809999998 podStartE2EDuration="2.447383981s" podCreationTimestamp="2025-10-02 09:42:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:43:01.445664337 +0000 UTC m=+369.093349564" watchObservedRunningTime="2025-10-02 09:43:01.447383981 +0000 UTC m=+369.095069048" Oct 02 09:43:02 crc kubenswrapper[4771]: I1002 09:43:02.426751 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-htfz2" event={"ID":"e302e18b-9d3f-4332-9ef9-37ac29846033","Type":"ContainerStarted","Data":"4bc0d6736e28ba68fb3552f4164cd48472f7118bcdc2aa11327537f40c62f1a6"} Oct 02 09:43:02 crc kubenswrapper[4771]: I1002 09:43:02.442396 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-htfz2" podStartSLOduration=2.089585944 podStartE2EDuration="3.442374809s" podCreationTimestamp="2025-10-02 09:42:59 +0000 UTC" firstStartedPulling="2025-10-02 09:43:00.450245148 +0000 UTC m=+368.097930215" lastFinishedPulling="2025-10-02 09:43:01.803034023 +0000 UTC m=+369.450719080" observedRunningTime="2025-10-02 09:43:02.441357262 +0000 UTC m=+370.089042349" watchObservedRunningTime="2025-10-02 09:43:02.442374809 +0000 UTC m=+370.090059876" Oct 02 09:43:03 crc kubenswrapper[4771]: I1002 09:43:03.433031 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-htfz2" Oct 02 09:43:03 crc kubenswrapper[4771]: I1002 09:43:03.438264 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-htfz2" Oct 02 09:43:03 crc kubenswrapper[4771]: I1002 09:43:03.919597 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-4b47f"] Oct 02 09:43:03 crc kubenswrapper[4771]: I1002 09:43:03.921050 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:03 crc kubenswrapper[4771]: I1002 09:43:03.925243 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Oct 02 09:43:03 crc kubenswrapper[4771]: I1002 09:43:03.925382 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-n5xrn" Oct 02 09:43:03 crc kubenswrapper[4771]: I1002 09:43:03.926425 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Oct 02 09:43:03 crc kubenswrapper[4771]: I1002 09:43:03.926646 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Oct 02 09:43:03 crc kubenswrapper[4771]: I1002 09:43:03.930670 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-4b47f"] Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.004063 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/491e0068-991b-41ea-adff-97d618266563-metrics-client-ca\") pod \"prometheus-operator-db54df47d-4b47f\" (UID: \"491e0068-991b-41ea-adff-97d618266563\") " pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.004112 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-267xw\" (UniqueName: \"kubernetes.io/projected/491e0068-991b-41ea-adff-97d618266563-kube-api-access-267xw\") pod \"prometheus-operator-db54df47d-4b47f\" (UID: \"491e0068-991b-41ea-adff-97d618266563\") " pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.004161 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/491e0068-991b-41ea-adff-97d618266563-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-4b47f\" (UID: \"491e0068-991b-41ea-adff-97d618266563\") " pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.004193 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/491e0068-991b-41ea-adff-97d618266563-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-4b47f\" (UID: \"491e0068-991b-41ea-adff-97d618266563\") " pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.105450 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-267xw\" (UniqueName: \"kubernetes.io/projected/491e0068-991b-41ea-adff-97d618266563-kube-api-access-267xw\") pod \"prometheus-operator-db54df47d-4b47f\" (UID: \"491e0068-991b-41ea-adff-97d618266563\") " pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.105508 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/491e0068-991b-41ea-adff-97d618266563-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-4b47f\" (UID: \"491e0068-991b-41ea-adff-97d618266563\") " pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.105535 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/491e0068-991b-41ea-adff-97d618266563-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-4b47f\" (UID: \"491e0068-991b-41ea-adff-97d618266563\") " pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.105589 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/491e0068-991b-41ea-adff-97d618266563-metrics-client-ca\") pod \"prometheus-operator-db54df47d-4b47f\" (UID: \"491e0068-991b-41ea-adff-97d618266563\") " pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.106706 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/491e0068-991b-41ea-adff-97d618266563-metrics-client-ca\") pod \"prometheus-operator-db54df47d-4b47f\" (UID: \"491e0068-991b-41ea-adff-97d618266563\") " pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.111373 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/491e0068-991b-41ea-adff-97d618266563-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-4b47f\" (UID: \"491e0068-991b-41ea-adff-97d618266563\") " pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.111372 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/491e0068-991b-41ea-adff-97d618266563-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-4b47f\" (UID: \"491e0068-991b-41ea-adff-97d618266563\") " pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.122982 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-267xw\" (UniqueName: \"kubernetes.io/projected/491e0068-991b-41ea-adff-97d618266563-kube-api-access-267xw\") pod \"prometheus-operator-db54df47d-4b47f\" (UID: \"491e0068-991b-41ea-adff-97d618266563\") " pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.238422 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.402548 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-4b47f"] Oct 02 09:43:04 crc kubenswrapper[4771]: W1002 09:43:04.412088 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod491e0068_991b_41ea_adff_97d618266563.slice/crio-6cb2d814fd46c196687a36eba6793141761c219478f6c95d357e71c75e4aaadb WatchSource:0}: Error finding container 6cb2d814fd46c196687a36eba6793141761c219478f6c95d357e71c75e4aaadb: Status 404 returned error can't find the container with id 6cb2d814fd46c196687a36eba6793141761c219478f6c95d357e71c75e4aaadb Oct 02 09:43:04 crc kubenswrapper[4771]: I1002 09:43:04.439906 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" event={"ID":"491e0068-991b-41ea-adff-97d618266563","Type":"ContainerStarted","Data":"6cb2d814fd46c196687a36eba6793141761c219478f6c95d357e71c75e4aaadb"} Oct 02 09:43:06 crc kubenswrapper[4771]: I1002 09:43:06.450723 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" event={"ID":"491e0068-991b-41ea-adff-97d618266563","Type":"ContainerStarted","Data":"e19b0803405f120e793f38b36421c3a2794c456ff71cfe78223dcf14dc2adb96"} Oct 02 09:43:06 crc kubenswrapper[4771]: I1002 09:43:06.451036 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" event={"ID":"491e0068-991b-41ea-adff-97d618266563","Type":"ContainerStarted","Data":"3c1a1e923d0ecdbed4631165bb5f22d27d48b53fa87168d4db0293436999c067"} Oct 02 09:43:06 crc kubenswrapper[4771]: I1002 09:43:06.468896 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-db54df47d-4b47f" podStartSLOduration=2.163971653 podStartE2EDuration="3.468880295s" podCreationTimestamp="2025-10-02 09:43:03 +0000 UTC" firstStartedPulling="2025-10-02 09:43:04.414428487 +0000 UTC m=+372.062113554" lastFinishedPulling="2025-10-02 09:43:05.719337129 +0000 UTC m=+373.367022196" observedRunningTime="2025-10-02 09:43:06.467631713 +0000 UTC m=+374.115316780" watchObservedRunningTime="2025-10-02 09:43:06.468880295 +0000 UTC m=+374.116565362" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.260531 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-j5n4v"] Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.261807 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.265061 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-vdnwv" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.265046 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.265980 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.271064 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-447wj"] Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.272437 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.276215 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.276479 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.278336 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-w55cw" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.282039 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-447wj"] Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.293020 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr"] Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.296435 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.297208 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr"] Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.298339 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.299586 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.300074 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.300364 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-2p6rh" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363600 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2eed3f65-86da-4591-8950-e509bbf98089-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363668 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-447wj\" (UID: \"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363697 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56vvb\" (UniqueName: \"kubernetes.io/projected/2eed3f65-86da-4591-8950-e509bbf98089-kube-api-access-56vvb\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363719 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-447wj\" (UID: \"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363737 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2eed3f65-86da-4591-8950-e509bbf98089-metrics-client-ca\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363759 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/2eed3f65-86da-4591-8950-e509bbf98089-node-exporter-textfile\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363780 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-447wj\" (UID: \"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363795 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363823 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/b82e224e-2444-4164-8201-cbe2fa39eee5-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363840 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-447tv\" (UniqueName: \"kubernetes.io/projected/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-api-access-447tv\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363856 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2eed3f65-86da-4591-8950-e509bbf98089-sys\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363874 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/2eed3f65-86da-4591-8950-e509bbf98089-node-exporter-wtmp\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363900 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcxwg\" (UniqueName: \"kubernetes.io/projected/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-kube-api-access-vcxwg\") pod \"openshift-state-metrics-566fddb674-447wj\" (UID: \"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363917 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b82e224e-2444-4164-8201-cbe2fa39eee5-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363941 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363969 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.363988 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/2eed3f65-86da-4591-8950-e509bbf98089-node-exporter-tls\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.364005 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/2eed3f65-86da-4591-8950-e509bbf98089-root\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.464970 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56vvb\" (UniqueName: \"kubernetes.io/projected/2eed3f65-86da-4591-8950-e509bbf98089-kube-api-access-56vvb\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465042 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-447wj\" (UID: \"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465074 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2eed3f65-86da-4591-8950-e509bbf98089-metrics-client-ca\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465113 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/2eed3f65-86da-4591-8950-e509bbf98089-node-exporter-textfile\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465162 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-447wj\" (UID: \"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465192 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465256 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/b82e224e-2444-4164-8201-cbe2fa39eee5-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465281 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2eed3f65-86da-4591-8950-e509bbf98089-sys\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465301 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-447tv\" (UniqueName: \"kubernetes.io/projected/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-api-access-447tv\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465324 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/2eed3f65-86da-4591-8950-e509bbf98089-node-exporter-wtmp\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465361 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcxwg\" (UniqueName: \"kubernetes.io/projected/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-kube-api-access-vcxwg\") pod \"openshift-state-metrics-566fddb674-447wj\" (UID: \"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465385 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b82e224e-2444-4164-8201-cbe2fa39eee5-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465419 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465458 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465489 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/2eed3f65-86da-4591-8950-e509bbf98089-node-exporter-tls\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465515 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/2eed3f65-86da-4591-8950-e509bbf98089-root\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465539 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2eed3f65-86da-4591-8950-e509bbf98089-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465566 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-447wj\" (UID: \"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:08 crc kubenswrapper[4771]: E1002 09:43:08.465693 4771 secret.go:188] Couldn't get secret openshift-monitoring/openshift-state-metrics-tls: secret "openshift-state-metrics-tls" not found Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465706 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2eed3f65-86da-4591-8950-e509bbf98089-sys\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: E1002 09:43:08.465748 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-openshift-state-metrics-tls podName:ca5af0a9-b5de-4bb8-afdb-87c2a64eb023 nodeName:}" failed. No retries permitted until 2025-10-02 09:43:08.965726486 +0000 UTC m=+376.613411553 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openshift-state-metrics-tls" (UniqueName: "kubernetes.io/secret/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-openshift-state-metrics-tls") pod "openshift-state-metrics-566fddb674-447wj" (UID: "ca5af0a9-b5de-4bb8-afdb-87c2a64eb023") : secret "openshift-state-metrics-tls" not found Oct 02 09:43:08 crc kubenswrapper[4771]: E1002 09:43:08.465805 4771 secret.go:188] Couldn't get secret openshift-monitoring/kube-state-metrics-tls: secret "kube-state-metrics-tls" not found Oct 02 09:43:08 crc kubenswrapper[4771]: E1002 09:43:08.465834 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-state-metrics-tls podName:b82e224e-2444-4164-8201-cbe2fa39eee5 nodeName:}" failed. No retries permitted until 2025-10-02 09:43:08.965822829 +0000 UTC m=+376.613507906 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-state-metrics-tls" (UniqueName: "kubernetes.io/secret/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-state-metrics-tls") pod "kube-state-metrics-777cb5bd5d-x52lr" (UID: "b82e224e-2444-4164-8201-cbe2fa39eee5") : secret "kube-state-metrics-tls" not found Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465858 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/2eed3f65-86da-4591-8950-e509bbf98089-node-exporter-textfile\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.465899 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/b82e224e-2444-4164-8201-cbe2fa39eee5-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.466629 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-447wj\" (UID: \"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.466635 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/2eed3f65-86da-4591-8950-e509bbf98089-node-exporter-wtmp\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.466684 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/2eed3f65-86da-4591-8950-e509bbf98089-root\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.466831 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.467529 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2eed3f65-86da-4591-8950-e509bbf98089-metrics-client-ca\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.467997 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b82e224e-2444-4164-8201-cbe2fa39eee5-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.476947 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2eed3f65-86da-4591-8950-e509bbf98089-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.477036 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-447wj\" (UID: \"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.476968 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.481114 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/2eed3f65-86da-4591-8950-e509bbf98089-node-exporter-tls\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.488852 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-447tv\" (UniqueName: \"kubernetes.io/projected/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-api-access-447tv\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.490738 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcxwg\" (UniqueName: \"kubernetes.io/projected/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-kube-api-access-vcxwg\") pod \"openshift-state-metrics-566fddb674-447wj\" (UID: \"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.492649 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56vvb\" (UniqueName: \"kubernetes.io/projected/2eed3f65-86da-4591-8950-e509bbf98089-kube-api-access-56vvb\") pod \"node-exporter-j5n4v\" (UID: \"2eed3f65-86da-4591-8950-e509bbf98089\") " pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.576811 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-j5n4v" Oct 02 09:43:08 crc kubenswrapper[4771]: W1002 09:43:08.598647 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2eed3f65_86da_4591_8950_e509bbf98089.slice/crio-2b5437d3c21febabf6e0cdc33807747f1912dbe58ef348ade7eb008a4b2c2acd WatchSource:0}: Error finding container 2b5437d3c21febabf6e0cdc33807747f1912dbe58ef348ade7eb008a4b2c2acd: Status 404 returned error can't find the container with id 2b5437d3c21febabf6e0cdc33807747f1912dbe58ef348ade7eb008a4b2c2acd Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.972068 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.972261 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-447wj\" (UID: \"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.976771 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/b82e224e-2444-4164-8201-cbe2fa39eee5-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-x52lr\" (UID: \"b82e224e-2444-4164-8201-cbe2fa39eee5\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:08 crc kubenswrapper[4771]: I1002 09:43:08.983967 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/ca5af0a9-b5de-4bb8-afdb-87c2a64eb023-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-447wj\" (UID: \"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.206891 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.227062 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.420540 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.422930 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.437172 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.437680 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-web-config" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.437708 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-dockercfg-bjlhf" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.438094 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-metric" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.438255 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls-assets-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.438301 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.438430 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-web" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.438657 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-generated" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.448627 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"alertmanager-trusted-ca-bundle" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.455002 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.478318 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/48b18937-6573-4f94-951e-c07317e17826-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.478371 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/48b18937-6573-4f94-951e-c07317e17826-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.478403 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.478452 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/48b18937-6573-4f94-951e-c07317e17826-config-out\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.478513 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rgz4\" (UniqueName: \"kubernetes.io/projected/48b18937-6573-4f94-951e-c07317e17826-kube-api-access-4rgz4\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.479051 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.479083 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/48b18937-6573-4f94-951e-c07317e17826-tls-assets\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.479226 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.479365 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-config-volume\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.479458 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/48b18937-6573-4f94-951e-c07317e17826-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.479488 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-web-config\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.479572 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.502510 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-j5n4v" event={"ID":"2eed3f65-86da-4591-8950-e509bbf98089","Type":"ContainerStarted","Data":"2b5437d3c21febabf6e0cdc33807747f1912dbe58ef348ade7eb008a4b2c2acd"} Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.581081 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rgz4\" (UniqueName: \"kubernetes.io/projected/48b18937-6573-4f94-951e-c07317e17826-kube-api-access-4rgz4\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.581145 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.581169 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/48b18937-6573-4f94-951e-c07317e17826-tls-assets\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.581193 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.581213 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-config-volume\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.581235 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/48b18937-6573-4f94-951e-c07317e17826-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.581249 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-web-config\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.581283 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.581310 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/48b18937-6573-4f94-951e-c07317e17826-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.581324 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/48b18937-6573-4f94-951e-c07317e17826-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.581343 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.581368 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/48b18937-6573-4f94-951e-c07317e17826-config-out\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.589143 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/48b18937-6573-4f94-951e-c07317e17826-config-out\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.590654 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/48b18937-6573-4f94-951e-c07317e17826-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.593141 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/48b18937-6573-4f94-951e-c07317e17826-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.597181 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rgz4\" (UniqueName: \"kubernetes.io/projected/48b18937-6573-4f94-951e-c07317e17826-kube-api-access-4rgz4\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.598724 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-web-config\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.598786 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.599280 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/48b18937-6573-4f94-951e-c07317e17826-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.599501 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.602776 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.604808 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/48b18937-6573-4f94-951e-c07317e17826-tls-assets\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.605197 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-config-volume\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.608768 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/48b18937-6573-4f94-951e-c07317e17826-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"48b18937-6573-4f94-951e-c07317e17826\") " pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.741071 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-447wj"] Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.771193 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Oct 02 09:43:09 crc kubenswrapper[4771]: I1002 09:43:09.803614 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr"] Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.230335 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-dcc5945f9-rzb7t"] Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.233021 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.236599 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.236665 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-rules" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.236736 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-web" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.236936 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-tls" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.237193 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-grpc-tls-3tb59kfs26qhl" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.237368 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-dockercfg-69q9k" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.239461 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-metrics" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.255187 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-dcc5945f9-rzb7t"] Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.295111 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.295189 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.295233 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-metrics-client-ca\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.295265 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-grpc-tls\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.295291 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.295328 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.295384 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjtl2\" (UniqueName: \"kubernetes.io/projected/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-kube-api-access-cjtl2\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.295440 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-tls\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.371069 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Oct 02 09:43:10 crc kubenswrapper[4771]: W1002 09:43:10.374973 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48b18937_6573_4f94_951e_c07317e17826.slice/crio-303148240f4bfc25dc3d0416467adb42645422bd927bd8f3eb2e015fa8957e9b WatchSource:0}: Error finding container 303148240f4bfc25dc3d0416467adb42645422bd927bd8f3eb2e015fa8957e9b: Status 404 returned error can't find the container with id 303148240f4bfc25dc3d0416467adb42645422bd927bd8f3eb2e015fa8957e9b Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.396984 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-tls\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.397056 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.397082 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.397107 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-metrics-client-ca\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.397143 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-grpc-tls\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.397163 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.397188 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.397222 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjtl2\" (UniqueName: \"kubernetes.io/projected/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-kube-api-access-cjtl2\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.398280 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-metrics-client-ca\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.403498 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.403557 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.408270 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.408963 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-grpc-tls\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.414411 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.423491 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-secret-thanos-querier-tls\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.423560 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjtl2\" (UniqueName: \"kubernetes.io/projected/b8b72cbe-278e-49c3-ac29-4586e20fc5c4-kube-api-access-cjtl2\") pod \"thanos-querier-dcc5945f9-rzb7t\" (UID: \"b8b72cbe-278e-49c3-ac29-4586e20fc5c4\") " pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.508239 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-j5n4v" event={"ID":"2eed3f65-86da-4591-8950-e509bbf98089","Type":"ContainerDied","Data":"633ca1d1806bda91839e28f0cc79a7ae8ab1df0332b25f63f7e3c584d8f76547"} Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.508106 4771 generic.go:334] "Generic (PLEG): container finished" podID="2eed3f65-86da-4591-8950-e509bbf98089" containerID="633ca1d1806bda91839e28f0cc79a7ae8ab1df0332b25f63f7e3c584d8f76547" exitCode=0 Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.514394 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"48b18937-6573-4f94-951e-c07317e17826","Type":"ContainerStarted","Data":"303148240f4bfc25dc3d0416467adb42645422bd927bd8f3eb2e015fa8957e9b"} Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.515940 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" event={"ID":"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023","Type":"ContainerStarted","Data":"5fbfcc672751fe3a773d71b19c2be218980d92cb980f3f068646cf1cc54dcaab"} Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.515967 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" event={"ID":"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023","Type":"ContainerStarted","Data":"bbde1477938843efa3d863cd6b118086c2bbeaa3de3e74ce50a84aef69a6b552"} Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.515978 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" event={"ID":"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023","Type":"ContainerStarted","Data":"89bbc80c6480b772afd94348d0baf6cef09dd0cc6161f5c45b00d9c34e3e610d"} Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.517026 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" event={"ID":"b82e224e-2444-4164-8201-cbe2fa39eee5","Type":"ContainerStarted","Data":"b8df2a7e8d152a2867a5a91defd7328e58049b5c3e9d4dc7c208e304c2f04b6c"} Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.556454 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:10 crc kubenswrapper[4771]: I1002 09:43:10.966002 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-dcc5945f9-rzb7t"] Oct 02 09:43:11 crc kubenswrapper[4771]: I1002 09:43:11.526372 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" event={"ID":"b8b72cbe-278e-49c3-ac29-4586e20fc5c4","Type":"ContainerStarted","Data":"147a437e29d6afff62f289e3becb68107ac3f44704ac2884e7ed1154eaa34e0c"} Oct 02 09:43:11 crc kubenswrapper[4771]: I1002 09:43:11.529502 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-j5n4v" event={"ID":"2eed3f65-86da-4591-8950-e509bbf98089","Type":"ContainerStarted","Data":"79fe79733a010d794472b77fabe746aa79a2939cbef7499a534a8b02dc0bfb3a"} Oct 02 09:43:11 crc kubenswrapper[4771]: I1002 09:43:11.529538 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-j5n4v" event={"ID":"2eed3f65-86da-4591-8950-e509bbf98089","Type":"ContainerStarted","Data":"55fc5d468fe53f9951ec188f7998b27bd60522ab2404649262539f27aa270fa8"} Oct 02 09:43:11 crc kubenswrapper[4771]: I1002 09:43:11.549724 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-j5n4v" podStartSLOduration=2.160156758 podStartE2EDuration="3.549705567s" podCreationTimestamp="2025-10-02 09:43:08 +0000 UTC" firstStartedPulling="2025-10-02 09:43:08.60108394 +0000 UTC m=+376.248769017" lastFinishedPulling="2025-10-02 09:43:09.990632759 +0000 UTC m=+377.638317826" observedRunningTime="2025-10-02 09:43:11.549249705 +0000 UTC m=+379.196934772" watchObservedRunningTime="2025-10-02 09:43:11.549705567 +0000 UTC m=+379.197390634" Oct 02 09:43:12 crc kubenswrapper[4771]: I1002 09:43:12.146354 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:43:12 crc kubenswrapper[4771]: I1002 09:43:12.146693 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:43:12 crc kubenswrapper[4771]: I1002 09:43:12.538695 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" event={"ID":"b82e224e-2444-4164-8201-cbe2fa39eee5","Type":"ContainerStarted","Data":"d7bfa7589f0b10f1847f7b3f38182a03607f34bf6563f74e3d1ae67aaf0233f1"} Oct 02 09:43:12 crc kubenswrapper[4771]: I1002 09:43:12.538756 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" event={"ID":"b82e224e-2444-4164-8201-cbe2fa39eee5","Type":"ContainerStarted","Data":"4327732296a238200da8956dad88572916b394c675d225a89152d06f63203eb6"} Oct 02 09:43:12 crc kubenswrapper[4771]: I1002 09:43:12.538771 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" event={"ID":"b82e224e-2444-4164-8201-cbe2fa39eee5","Type":"ContainerStarted","Data":"1833035ad76251253fdbe0b4bd4e5c85e67c60ac52d6adbf40be0f63cbde5bff"} Oct 02 09:43:12 crc kubenswrapper[4771]: I1002 09:43:12.543190 4771 generic.go:334] "Generic (PLEG): container finished" podID="48b18937-6573-4f94-951e-c07317e17826" containerID="0102a72fbab2ca5e3ca40a30f7a59c512a9aab3bf494c5d172758ac2db689a68" exitCode=0 Oct 02 09:43:12 crc kubenswrapper[4771]: I1002 09:43:12.543254 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"48b18937-6573-4f94-951e-c07317e17826","Type":"ContainerDied","Data":"0102a72fbab2ca5e3ca40a30f7a59c512a9aab3bf494c5d172758ac2db689a68"} Oct 02 09:43:12 crc kubenswrapper[4771]: I1002 09:43:12.548864 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" event={"ID":"ca5af0a9-b5de-4bb8-afdb-87c2a64eb023","Type":"ContainerStarted","Data":"bac33a0841c4030eb7e108658bba2cb913ed54b0237d7f19311bee743e9402a9"} Oct 02 09:43:12 crc kubenswrapper[4771]: I1002 09:43:12.559812 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-x52lr" podStartSLOduration=2.556210959 podStartE2EDuration="4.559796965s" podCreationTimestamp="2025-10-02 09:43:08 +0000 UTC" firstStartedPulling="2025-10-02 09:43:09.94600159 +0000 UTC m=+377.593686657" lastFinishedPulling="2025-10-02 09:43:11.949587596 +0000 UTC m=+379.597272663" observedRunningTime="2025-10-02 09:43:12.558113631 +0000 UTC m=+380.205798708" watchObservedRunningTime="2025-10-02 09:43:12.559796965 +0000 UTC m=+380.207482032" Oct 02 09:43:12 crc kubenswrapper[4771]: I1002 09:43:12.577716 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-566fddb674-447wj" podStartSLOduration=2.889356725 podStartE2EDuration="4.577700319s" podCreationTimestamp="2025-10-02 09:43:08 +0000 UTC" firstStartedPulling="2025-10-02 09:43:10.261090898 +0000 UTC m=+377.908775965" lastFinishedPulling="2025-10-02 09:43:11.949434492 +0000 UTC m=+379.597119559" observedRunningTime="2025-10-02 09:43:12.573559152 +0000 UTC m=+380.221244229" watchObservedRunningTime="2025-10-02 09:43:12.577700319 +0000 UTC m=+380.225385386" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.099332 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-698d48b8b5-zhfsf"] Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.106345 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.107289 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-698d48b8b5-zhfsf"] Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.141206 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9f32682-53d4-4987-9402-265115bc84df-console-serving-cert\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.141285 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt4kn\" (UniqueName: \"kubernetes.io/projected/a9f32682-53d4-4987-9402-265115bc84df-kube-api-access-mt4kn\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.141312 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-service-ca\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.141348 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a9f32682-53d4-4987-9402-265115bc84df-console-oauth-config\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.141375 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-oauth-serving-cert\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.141408 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-trusted-ca-bundle\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.141543 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-console-config\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.243004 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-service-ca\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.243072 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a9f32682-53d4-4987-9402-265115bc84df-console-oauth-config\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.243102 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-oauth-serving-cert\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.243154 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-trusted-ca-bundle\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.243197 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-console-config\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.243253 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9f32682-53d4-4987-9402-265115bc84df-console-serving-cert\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.243287 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt4kn\" (UniqueName: \"kubernetes.io/projected/a9f32682-53d4-4987-9402-265115bc84df-kube-api-access-mt4kn\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.244510 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-console-config\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.244629 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-oauth-serving-cert\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.245239 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-service-ca\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.245506 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-trusted-ca-bundle\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.250532 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9f32682-53d4-4987-9402-265115bc84df-console-serving-cert\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.260400 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a9f32682-53d4-4987-9402-265115bc84df-console-oauth-config\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.261333 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt4kn\" (UniqueName: \"kubernetes.io/projected/a9f32682-53d4-4987-9402-265115bc84df-kube-api-access-mt4kn\") pod \"console-698d48b8b5-zhfsf\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.423964 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.670024 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-6cd7d74865-xhd2p"] Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.672927 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.675881 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-6cd7d74865-xhd2p"] Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.678015 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.678281 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-k1is2r970shn" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.678513 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.678760 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.678987 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-b9c77" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.679328 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.851376 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9ef9b81-e70b-4adf-bd83-57e727084f73-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.851763 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a9ef9b81-e70b-4adf-bd83-57e727084f73-audit-log\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.851804 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/a9ef9b81-e70b-4adf-bd83-57e727084f73-secret-metrics-client-certs\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.851828 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a9ef9b81-e70b-4adf-bd83-57e727084f73-metrics-server-audit-profiles\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.851852 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qxlp\" (UniqueName: \"kubernetes.io/projected/a9ef9b81-e70b-4adf-bd83-57e727084f73-kube-api-access-5qxlp\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.851968 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ef9b81-e70b-4adf-bd83-57e727084f73-client-ca-bundle\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.852016 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/a9ef9b81-e70b-4adf-bd83-57e727084f73-secret-metrics-server-tls\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.952998 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/a9ef9b81-e70b-4adf-bd83-57e727084f73-secret-metrics-server-tls\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.953078 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9ef9b81-e70b-4adf-bd83-57e727084f73-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.953103 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a9ef9b81-e70b-4adf-bd83-57e727084f73-audit-log\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.953178 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/a9ef9b81-e70b-4adf-bd83-57e727084f73-secret-metrics-client-certs\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.953207 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a9ef9b81-e70b-4adf-bd83-57e727084f73-metrics-server-audit-profiles\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.953234 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qxlp\" (UniqueName: \"kubernetes.io/projected/a9ef9b81-e70b-4adf-bd83-57e727084f73-kube-api-access-5qxlp\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.953296 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ef9b81-e70b-4adf-bd83-57e727084f73-client-ca-bundle\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.953960 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a9ef9b81-e70b-4adf-bd83-57e727084f73-audit-log\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.955235 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a9ef9b81-e70b-4adf-bd83-57e727084f73-metrics-server-audit-profiles\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.955933 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9ef9b81-e70b-4adf-bd83-57e727084f73-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.963264 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/a9ef9b81-e70b-4adf-bd83-57e727084f73-secret-metrics-server-tls\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.963395 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/a9ef9b81-e70b-4adf-bd83-57e727084f73-secret-metrics-client-certs\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.963554 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ef9b81-e70b-4adf-bd83-57e727084f73-client-ca-bundle\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.970965 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qxlp\" (UniqueName: \"kubernetes.io/projected/a9ef9b81-e70b-4adf-bd83-57e727084f73-kube-api-access-5qxlp\") pod \"metrics-server-6cd7d74865-xhd2p\" (UID: \"a9ef9b81-e70b-4adf-bd83-57e727084f73\") " pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:13 crc kubenswrapper[4771]: I1002 09:43:13.996112 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.070873 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-77d464cd8b-8wpg6"] Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.071880 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-77d464cd8b-8wpg6" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.075065 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-6tstp" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.075355 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.086451 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-77d464cd8b-8wpg6"] Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.136248 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-698d48b8b5-zhfsf"] Oct 02 09:43:14 crc kubenswrapper[4771]: W1002 09:43:14.152684 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9f32682_53d4_4987_9402_265115bc84df.slice/crio-9d92fdc8d41d156779a1f61d0f6624c3a6830aba23ebcd1a98cd8fd104fb08b0 WatchSource:0}: Error finding container 9d92fdc8d41d156779a1f61d0f6624c3a6830aba23ebcd1a98cd8fd104fb08b0: Status 404 returned error can't find the container with id 9d92fdc8d41d156779a1f61d0f6624c3a6830aba23ebcd1a98cd8fd104fb08b0 Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.155323 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/7da1f6ae-0271-46b8-aa4e-e31ef38053ea-monitoring-plugin-cert\") pod \"monitoring-plugin-77d464cd8b-8wpg6\" (UID: \"7da1f6ae-0271-46b8-aa4e-e31ef38053ea\") " pod="openshift-monitoring/monitoring-plugin-77d464cd8b-8wpg6" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.256310 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/7da1f6ae-0271-46b8-aa4e-e31ef38053ea-monitoring-plugin-cert\") pod \"monitoring-plugin-77d464cd8b-8wpg6\" (UID: \"7da1f6ae-0271-46b8-aa4e-e31ef38053ea\") " pod="openshift-monitoring/monitoring-plugin-77d464cd8b-8wpg6" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.263733 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/7da1f6ae-0271-46b8-aa4e-e31ef38053ea-monitoring-plugin-cert\") pod \"monitoring-plugin-77d464cd8b-8wpg6\" (UID: \"7da1f6ae-0271-46b8-aa4e-e31ef38053ea\") " pod="openshift-monitoring/monitoring-plugin-77d464cd8b-8wpg6" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.414495 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-6cd7d74865-xhd2p"] Oct 02 09:43:14 crc kubenswrapper[4771]: W1002 09:43:14.420797 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9ef9b81_e70b_4adf_bd83_57e727084f73.slice/crio-45cee06fd47ecf7780848561cdbb328bf0630f42d6947b08eed1d303fa380d19 WatchSource:0}: Error finding container 45cee06fd47ecf7780848561cdbb328bf0630f42d6947b08eed1d303fa380d19: Status 404 returned error can't find the container with id 45cee06fd47ecf7780848561cdbb328bf0630f42d6947b08eed1d303fa380d19 Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.429024 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-77d464cd8b-8wpg6" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.571185 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" event={"ID":"a9ef9b81-e70b-4adf-bd83-57e727084f73","Type":"ContainerStarted","Data":"45cee06fd47ecf7780848561cdbb328bf0630f42d6947b08eed1d303fa380d19"} Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.578412 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" event={"ID":"b8b72cbe-278e-49c3-ac29-4586e20fc5c4","Type":"ContainerStarted","Data":"25892a17a97c7dd4decb294960d3a07e7da9633a46b6ab908bc0d600f65b9fad"} Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.578457 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" event={"ID":"b8b72cbe-278e-49c3-ac29-4586e20fc5c4","Type":"ContainerStarted","Data":"1f09f254c2f21b3f4cb8c62d911627c3c7dc912233eded1686b08201b61eef4a"} Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.578466 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" event={"ID":"b8b72cbe-278e-49c3-ac29-4586e20fc5c4","Type":"ContainerStarted","Data":"53ed5bb71a6281f2917d7e93fb62d3ddc4e3b4a949ec5c4d94d110d764bef4c2"} Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.581254 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-698d48b8b5-zhfsf" event={"ID":"a9f32682-53d4-4987-9402-265115bc84df","Type":"ContainerStarted","Data":"616f6c98734dc2ad5642a6567d01fb31f5bbf884f25f7e3b3217a42128f2b22d"} Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.581310 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-698d48b8b5-zhfsf" event={"ID":"a9f32682-53d4-4987-9402-265115bc84df","Type":"ContainerStarted","Data":"9d92fdc8d41d156779a1f61d0f6624c3a6830aba23ebcd1a98cd8fd104fb08b0"} Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.590978 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.593405 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.597619 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-rbac-proxy" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.598443 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"serving-certs-ca-bundle" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.598460 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-prometheus-http-client-file" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.598469 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-web-config" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.599274 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.620107 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-dockercfg-qpfqr" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.622810 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-sidecar-tls" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.623103 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls-assets-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.626197 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.626307 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-grpc-tls-4bjj2an6h4gj3" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.626325 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-kube-rbac-proxy-web" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.635583 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-698d48b8b5-zhfsf" podStartSLOduration=1.635567215 podStartE2EDuration="1.635567215s" podCreationTimestamp="2025-10-02 09:43:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:43:14.6338433 +0000 UTC m=+382.281528377" watchObservedRunningTime="2025-10-02 09:43:14.635567215 +0000 UTC m=+382.283252272" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.647423 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-k8s-rulefiles-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.647864 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-trusted-ca-bundle" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663383 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-config\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663426 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663455 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663473 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e32bc476-2cfa-4860-82df-f35c8bffa031-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663529 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/e32bc476-2cfa-4860-82df-f35c8bffa031-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663576 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663628 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663735 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd7gn\" (UniqueName: \"kubernetes.io/projected/e32bc476-2cfa-4860-82df-f35c8bffa031-kube-api-access-qd7gn\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663811 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663848 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663873 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663894 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-web-config\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663921 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.663981 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.664004 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.664033 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.664062 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e32bc476-2cfa-4860-82df-f35c8bffa031-config-out\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.664105 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.668146 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.690606 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-77d464cd8b-8wpg6"] Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.765478 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.765587 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-config\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.765615 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766364 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766439 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e32bc476-2cfa-4860-82df-f35c8bffa031-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766464 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/e32bc476-2cfa-4860-82df-f35c8bffa031-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766506 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766554 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766587 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd7gn\" (UniqueName: \"kubernetes.io/projected/e32bc476-2cfa-4860-82df-f35c8bffa031-kube-api-access-qd7gn\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766624 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766647 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766667 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766689 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-web-config\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766712 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766761 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766786 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766812 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766867 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.766879 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e32bc476-2cfa-4860-82df-f35c8bffa031-config-out\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.768846 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.769050 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.769423 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.771044 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/e32bc476-2cfa-4860-82df-f35c8bffa031-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.772317 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e32bc476-2cfa-4860-82df-f35c8bffa031-config-out\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.773177 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.773451 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.776012 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e32bc476-2cfa-4860-82df-f35c8bffa031-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.776901 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-config\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.776962 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.777847 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.778386 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.778645 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.778717 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.779155 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e32bc476-2cfa-4860-82df-f35c8bffa031-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.785417 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd7gn\" (UniqueName: \"kubernetes.io/projected/e32bc476-2cfa-4860-82df-f35c8bffa031-kube-api-access-qd7gn\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.787940 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e32bc476-2cfa-4860-82df-f35c8bffa031-web-config\") pod \"prometheus-k8s-0\" (UID: \"e32bc476-2cfa-4860-82df-f35c8bffa031\") " pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:14 crc kubenswrapper[4771]: I1002 09:43:14.939051 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:15 crc kubenswrapper[4771]: I1002 09:43:15.588566 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-77d464cd8b-8wpg6" event={"ID":"7da1f6ae-0271-46b8-aa4e-e31ef38053ea","Type":"ContainerStarted","Data":"67e11c2843864b9bc5ba4b18629ea7d17b35b5000a57f4e6347e63d5ea218d40"} Oct 02 09:43:15 crc kubenswrapper[4771]: I1002 09:43:15.742435 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Oct 02 09:43:15 crc kubenswrapper[4771]: W1002 09:43:15.749161 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode32bc476_2cfa_4860_82df_f35c8bffa031.slice/crio-ec6644e9ba76cfc6fafb9308f3c801a9d6e611f107a50bc5ac9b9685f2668649 WatchSource:0}: Error finding container ec6644e9ba76cfc6fafb9308f3c801a9d6e611f107a50bc5ac9b9685f2668649: Status 404 returned error can't find the container with id ec6644e9ba76cfc6fafb9308f3c801a9d6e611f107a50bc5ac9b9685f2668649 Oct 02 09:43:16 crc kubenswrapper[4771]: I1002 09:43:16.596921 4771 generic.go:334] "Generic (PLEG): container finished" podID="e32bc476-2cfa-4860-82df-f35c8bffa031" containerID="801b2a11d46b4ddbba7f7b69f4245cd7d68c777e648413531c3e4f6a425d46c8" exitCode=0 Oct 02 09:43:16 crc kubenswrapper[4771]: I1002 09:43:16.596976 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e32bc476-2cfa-4860-82df-f35c8bffa031","Type":"ContainerDied","Data":"801b2a11d46b4ddbba7f7b69f4245cd7d68c777e648413531c3e4f6a425d46c8"} Oct 02 09:43:16 crc kubenswrapper[4771]: I1002 09:43:16.597334 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e32bc476-2cfa-4860-82df-f35c8bffa031","Type":"ContainerStarted","Data":"ec6644e9ba76cfc6fafb9308f3c801a9d6e611f107a50bc5ac9b9685f2668649"} Oct 02 09:43:16 crc kubenswrapper[4771]: I1002 09:43:16.604248 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" event={"ID":"b8b72cbe-278e-49c3-ac29-4586e20fc5c4","Type":"ContainerStarted","Data":"4629e56950f4be87893a1eac59777fd80486576621fdee12fd63607222e3f064"} Oct 02 09:43:16 crc kubenswrapper[4771]: I1002 09:43:16.604289 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" event={"ID":"b8b72cbe-278e-49c3-ac29-4586e20fc5c4","Type":"ContainerStarted","Data":"ed97cd1c3396ec9a2e9f3b380b7f750959b0af19519239d126b5b5ee97546758"} Oct 02 09:43:16 crc kubenswrapper[4771]: I1002 09:43:16.604300 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" event={"ID":"b8b72cbe-278e-49c3-ac29-4586e20fc5c4","Type":"ContainerStarted","Data":"1a86694a330fe9f95a147757513b0bee843c22b49e42ba9457909fd144607027"} Oct 02 09:43:16 crc kubenswrapper[4771]: I1002 09:43:16.604897 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:16 crc kubenswrapper[4771]: I1002 09:43:16.607684 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"48b18937-6573-4f94-951e-c07317e17826","Type":"ContainerStarted","Data":"a08645bceb4ef2391acad2e4c718021d2c839c1fd47ebd3dca086e7704c0d5df"} Oct 02 09:43:16 crc kubenswrapper[4771]: I1002 09:43:16.607716 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"48b18937-6573-4f94-951e-c07317e17826","Type":"ContainerStarted","Data":"38c430279caa70628cb1a3ea5a07f65c8d8fb47175db21f58a3df5b91e6f7cdf"} Oct 02 09:43:16 crc kubenswrapper[4771]: I1002 09:43:16.607730 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"48b18937-6573-4f94-951e-c07317e17826","Type":"ContainerStarted","Data":"ddb63be1b085f7d8171f7018187d1be40a0de05ec1d2b7ee470378d69730849d"} Oct 02 09:43:16 crc kubenswrapper[4771]: I1002 09:43:16.662318 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" podStartSLOduration=2.296656085 podStartE2EDuration="6.662294532s" podCreationTimestamp="2025-10-02 09:43:10 +0000 UTC" firstStartedPulling="2025-10-02 09:43:11.284314628 +0000 UTC m=+378.931999705" lastFinishedPulling="2025-10-02 09:43:15.649953085 +0000 UTC m=+383.297638152" observedRunningTime="2025-10-02 09:43:16.653507794 +0000 UTC m=+384.301192871" watchObservedRunningTime="2025-10-02 09:43:16.662294532 +0000 UTC m=+384.309979599" Oct 02 09:43:17 crc kubenswrapper[4771]: I1002 09:43:17.617282 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" event={"ID":"a9ef9b81-e70b-4adf-bd83-57e727084f73","Type":"ContainerStarted","Data":"d63ebe52ab4d9762a8202922c7936457056afbae0eebbc107f13e16f88fff15a"} Oct 02 09:43:17 crc kubenswrapper[4771]: I1002 09:43:17.620644 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-77d464cd8b-8wpg6" event={"ID":"7da1f6ae-0271-46b8-aa4e-e31ef38053ea","Type":"ContainerStarted","Data":"f58a14948168c45c868146bc2464f7879e76571125a5f558ac93a2b864cfd216"} Oct 02 09:43:17 crc kubenswrapper[4771]: I1002 09:43:17.620925 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-77d464cd8b-8wpg6" Oct 02 09:43:17 crc kubenswrapper[4771]: I1002 09:43:17.625599 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"48b18937-6573-4f94-951e-c07317e17826","Type":"ContainerStarted","Data":"0125d77218df6f49b1b8ae8c531dd3034bc993b3facede22deddc909951e6b83"} Oct 02 09:43:17 crc kubenswrapper[4771]: I1002 09:43:17.625797 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"48b18937-6573-4f94-951e-c07317e17826","Type":"ContainerStarted","Data":"a383be1cdc53f251b588ea4597c1519011f3728dfe3be6740b31503b3cc29580"} Oct 02 09:43:17 crc kubenswrapper[4771]: I1002 09:43:17.625943 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"48b18937-6573-4f94-951e-c07317e17826","Type":"ContainerStarted","Data":"d3e1eebf106293d409b6757e9443e115ea45f6e91ef98213015008b8b7ce48d5"} Oct 02 09:43:17 crc kubenswrapper[4771]: I1002 09:43:17.626817 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-77d464cd8b-8wpg6" Oct 02 09:43:17 crc kubenswrapper[4771]: I1002 09:43:17.639920 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" podStartSLOduration=2.186475764 podStartE2EDuration="4.639899667s" podCreationTimestamp="2025-10-02 09:43:13 +0000 UTC" firstStartedPulling="2025-10-02 09:43:14.422869144 +0000 UTC m=+382.070554211" lastFinishedPulling="2025-10-02 09:43:16.876293047 +0000 UTC m=+384.523978114" observedRunningTime="2025-10-02 09:43:17.633782739 +0000 UTC m=+385.281467836" watchObservedRunningTime="2025-10-02 09:43:17.639899667 +0000 UTC m=+385.287584734" Oct 02 09:43:17 crc kubenswrapper[4771]: I1002 09:43:17.652737 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-77d464cd8b-8wpg6" podStartSLOduration=1.489127919 podStartE2EDuration="3.65271547s" podCreationTimestamp="2025-10-02 09:43:14 +0000 UTC" firstStartedPulling="2025-10-02 09:43:14.708566529 +0000 UTC m=+382.356251596" lastFinishedPulling="2025-10-02 09:43:16.87215408 +0000 UTC m=+384.519839147" observedRunningTime="2025-10-02 09:43:17.649885737 +0000 UTC m=+385.297570814" watchObservedRunningTime="2025-10-02 09:43:17.65271547 +0000 UTC m=+385.300400547" Oct 02 09:43:17 crc kubenswrapper[4771]: I1002 09:43:17.685948 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=3.445956321 podStartE2EDuration="8.685909062s" podCreationTimestamp="2025-10-02 09:43:09 +0000 UTC" firstStartedPulling="2025-10-02 09:43:10.377228713 +0000 UTC m=+378.024913780" lastFinishedPulling="2025-10-02 09:43:15.617181454 +0000 UTC m=+383.264866521" observedRunningTime="2025-10-02 09:43:17.676350484 +0000 UTC m=+385.324035551" watchObservedRunningTime="2025-10-02 09:43:17.685909062 +0000 UTC m=+385.333594129" Oct 02 09:43:20 crc kubenswrapper[4771]: I1002 09:43:20.066487 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-pq9fh" Oct 02 09:43:20 crc kubenswrapper[4771]: I1002 09:43:20.138510 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p2srt"] Oct 02 09:43:20 crc kubenswrapper[4771]: I1002 09:43:20.570521 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-dcc5945f9-rzb7t" Oct 02 09:43:20 crc kubenswrapper[4771]: I1002 09:43:20.655092 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e32bc476-2cfa-4860-82df-f35c8bffa031","Type":"ContainerStarted","Data":"fdfdcd16e05a3058ef0a4a3fa81b686efb8e85a75d94622fade56fb578180812"} Oct 02 09:43:20 crc kubenswrapper[4771]: I1002 09:43:20.655166 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e32bc476-2cfa-4860-82df-f35c8bffa031","Type":"ContainerStarted","Data":"b9101d607c314be0e6e46ad4d63e9e60d4f85100a957899d76b4acfb990f96a3"} Oct 02 09:43:20 crc kubenswrapper[4771]: I1002 09:43:20.655186 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e32bc476-2cfa-4860-82df-f35c8bffa031","Type":"ContainerStarted","Data":"2695ddc4577177fe10fc23712f9739bf1e50f4b8b65e5552f0d72fc16b8dbd25"} Oct 02 09:43:20 crc kubenswrapper[4771]: I1002 09:43:20.655198 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e32bc476-2cfa-4860-82df-f35c8bffa031","Type":"ContainerStarted","Data":"f4fc4a8a354a3716b50a45dbf1580e607b385959216f40168115011029c2cfac"} Oct 02 09:43:21 crc kubenswrapper[4771]: I1002 09:43:21.664421 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e32bc476-2cfa-4860-82df-f35c8bffa031","Type":"ContainerStarted","Data":"e1b958406e43c564018bf3925e35f52770ea686208bd4d19aac747c0116e8c48"} Oct 02 09:43:21 crc kubenswrapper[4771]: I1002 09:43:21.664745 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e32bc476-2cfa-4860-82df-f35c8bffa031","Type":"ContainerStarted","Data":"3b46914f78faf015c67cbc7540d5849dd33d1162cbfff2c029cbe3671294c2f4"} Oct 02 09:43:21 crc kubenswrapper[4771]: I1002 09:43:21.699684 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=4.288523661 podStartE2EDuration="7.699666512s" podCreationTimestamp="2025-10-02 09:43:14 +0000 UTC" firstStartedPulling="2025-10-02 09:43:16.598877226 +0000 UTC m=+384.246562283" lastFinishedPulling="2025-10-02 09:43:20.010020067 +0000 UTC m=+387.657705134" observedRunningTime="2025-10-02 09:43:21.699218831 +0000 UTC m=+389.346903898" watchObservedRunningTime="2025-10-02 09:43:21.699666512 +0000 UTC m=+389.347351589" Oct 02 09:43:23 crc kubenswrapper[4771]: I1002 09:43:23.424495 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:23 crc kubenswrapper[4771]: I1002 09:43:23.424883 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:23 crc kubenswrapper[4771]: I1002 09:43:23.430329 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:23 crc kubenswrapper[4771]: I1002 09:43:23.679096 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:43:23 crc kubenswrapper[4771]: I1002 09:43:23.734680 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-r664s"] Oct 02 09:43:24 crc kubenswrapper[4771]: I1002 09:43:24.940099 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:43:33 crc kubenswrapper[4771]: I1002 09:43:33.998199 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:33 crc kubenswrapper[4771]: I1002 09:43:33.999012 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:42 crc kubenswrapper[4771]: I1002 09:43:42.146190 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:43:42 crc kubenswrapper[4771]: I1002 09:43:42.147037 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:43:42 crc kubenswrapper[4771]: I1002 09:43:42.147084 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:43:42 crc kubenswrapper[4771]: I1002 09:43:42.147766 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"78b7a8d589783481aba6ddc2e3b7cc7710dc210299af82522a169351aa5ee352"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:43:42 crc kubenswrapper[4771]: I1002 09:43:42.147823 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://78b7a8d589783481aba6ddc2e3b7cc7710dc210299af82522a169351aa5ee352" gracePeriod=600 Oct 02 09:43:42 crc kubenswrapper[4771]: I1002 09:43:42.797922 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="78b7a8d589783481aba6ddc2e3b7cc7710dc210299af82522a169351aa5ee352" exitCode=0 Oct 02 09:43:42 crc kubenswrapper[4771]: I1002 09:43:42.797992 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"78b7a8d589783481aba6ddc2e3b7cc7710dc210299af82522a169351aa5ee352"} Oct 02 09:43:42 crc kubenswrapper[4771]: I1002 09:43:42.798636 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"f2d56a0106e4d97ef1eaacdf9670a1e221241c6eb398fb493028ecc6d5fd6e25"} Oct 02 09:43:42 crc kubenswrapper[4771]: I1002 09:43:42.798656 4771 scope.go:117] "RemoveContainer" containerID="5f717dc010c8ed68981cdc9d0b6315237bcc922bf69329a8a72d9394d7c8af3a" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.200269 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" podUID="675e6316-9501-4ac0-a134-de142b7bfcd4" containerName="registry" containerID="cri-o://bcf430f2779768e996dae1998ca3fc09a6a43d5ff5eac647ab3f05b175ed0c5c" gracePeriod=30 Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.533491 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.655285 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-bound-sa-token\") pod \"675e6316-9501-4ac0-a134-de142b7bfcd4\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.655557 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"675e6316-9501-4ac0-a134-de142b7bfcd4\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.655590 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/675e6316-9501-4ac0-a134-de142b7bfcd4-ca-trust-extracted\") pod \"675e6316-9501-4ac0-a134-de142b7bfcd4\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.655664 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/675e6316-9501-4ac0-a134-de142b7bfcd4-registry-certificates\") pod \"675e6316-9501-4ac0-a134-de142b7bfcd4\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.655720 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbjm4\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-kube-api-access-sbjm4\") pod \"675e6316-9501-4ac0-a134-de142b7bfcd4\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.655745 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-registry-tls\") pod \"675e6316-9501-4ac0-a134-de142b7bfcd4\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.655772 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/675e6316-9501-4ac0-a134-de142b7bfcd4-trusted-ca\") pod \"675e6316-9501-4ac0-a134-de142b7bfcd4\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.655818 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/675e6316-9501-4ac0-a134-de142b7bfcd4-installation-pull-secrets\") pod \"675e6316-9501-4ac0-a134-de142b7bfcd4\" (UID: \"675e6316-9501-4ac0-a134-de142b7bfcd4\") " Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.656937 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/675e6316-9501-4ac0-a134-de142b7bfcd4-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "675e6316-9501-4ac0-a134-de142b7bfcd4" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.658212 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/675e6316-9501-4ac0-a134-de142b7bfcd4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "675e6316-9501-4ac0-a134-de142b7bfcd4" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.663853 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "675e6316-9501-4ac0-a134-de142b7bfcd4" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.665876 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "675e6316-9501-4ac0-a134-de142b7bfcd4" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.665892 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-kube-api-access-sbjm4" (OuterVolumeSpecName: "kube-api-access-sbjm4") pod "675e6316-9501-4ac0-a134-de142b7bfcd4" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4"). InnerVolumeSpecName "kube-api-access-sbjm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.666057 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/675e6316-9501-4ac0-a134-de142b7bfcd4-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "675e6316-9501-4ac0-a134-de142b7bfcd4" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.666180 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "675e6316-9501-4ac0-a134-de142b7bfcd4" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.674321 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/675e6316-9501-4ac0-a134-de142b7bfcd4-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "675e6316-9501-4ac0-a134-de142b7bfcd4" (UID: "675e6316-9501-4ac0-a134-de142b7bfcd4"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.757677 4771 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/675e6316-9501-4ac0-a134-de142b7bfcd4-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.757714 4771 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.757724 4771 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/675e6316-9501-4ac0-a134-de142b7bfcd4-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.757733 4771 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/675e6316-9501-4ac0-a134-de142b7bfcd4-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.757742 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbjm4\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-kube-api-access-sbjm4\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.757750 4771 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/675e6316-9501-4ac0-a134-de142b7bfcd4-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.757759 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/675e6316-9501-4ac0-a134-de142b7bfcd4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.821800 4771 generic.go:334] "Generic (PLEG): container finished" podID="675e6316-9501-4ac0-a134-de142b7bfcd4" containerID="bcf430f2779768e996dae1998ca3fc09a6a43d5ff5eac647ab3f05b175ed0c5c" exitCode=0 Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.821846 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.821852 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" event={"ID":"675e6316-9501-4ac0-a134-de142b7bfcd4","Type":"ContainerDied","Data":"bcf430f2779768e996dae1998ca3fc09a6a43d5ff5eac647ab3f05b175ed0c5c"} Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.821881 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p2srt" event={"ID":"675e6316-9501-4ac0-a134-de142b7bfcd4","Type":"ContainerDied","Data":"a8d14d1a87916d0224041f958acc63022f79c22db3c8500c83af3329ec4b250c"} Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.821902 4771 scope.go:117] "RemoveContainer" containerID="bcf430f2779768e996dae1998ca3fc09a6a43d5ff5eac647ab3f05b175ed0c5c" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.849875 4771 scope.go:117] "RemoveContainer" containerID="bcf430f2779768e996dae1998ca3fc09a6a43d5ff5eac647ab3f05b175ed0c5c" Oct 02 09:43:45 crc kubenswrapper[4771]: E1002 09:43:45.850739 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcf430f2779768e996dae1998ca3fc09a6a43d5ff5eac647ab3f05b175ed0c5c\": container with ID starting with bcf430f2779768e996dae1998ca3fc09a6a43d5ff5eac647ab3f05b175ed0c5c not found: ID does not exist" containerID="bcf430f2779768e996dae1998ca3fc09a6a43d5ff5eac647ab3f05b175ed0c5c" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.850797 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcf430f2779768e996dae1998ca3fc09a6a43d5ff5eac647ab3f05b175ed0c5c"} err="failed to get container status \"bcf430f2779768e996dae1998ca3fc09a6a43d5ff5eac647ab3f05b175ed0c5c\": rpc error: code = NotFound desc = could not find container \"bcf430f2779768e996dae1998ca3fc09a6a43d5ff5eac647ab3f05b175ed0c5c\": container with ID starting with bcf430f2779768e996dae1998ca3fc09a6a43d5ff5eac647ab3f05b175ed0c5c not found: ID does not exist" Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.851044 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p2srt"] Oct 02 09:43:45 crc kubenswrapper[4771]: I1002 09:43:45.861004 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p2srt"] Oct 02 09:43:47 crc kubenswrapper[4771]: I1002 09:43:47.687647 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="675e6316-9501-4ac0-a134-de142b7bfcd4" path="/var/lib/kubelet/pods/675e6316-9501-4ac0-a134-de142b7bfcd4/volumes" Oct 02 09:43:48 crc kubenswrapper[4771]: I1002 09:43:48.772962 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-r664s" podUID="67d2ef8d-5a0c-48fa-8233-9bb23e794da0" containerName="console" containerID="cri-o://a118fa7267dd547ef56f471f662395be2a9720da4117eb6fef6014523b8ff519" gracePeriod=15 Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.134803 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-r664s_67d2ef8d-5a0c-48fa-8233-9bb23e794da0/console/0.log" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.135215 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.307439 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-oauth-serving-cert\") pod \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.307481 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-trusted-ca-bundle\") pod \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.307578 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-config\") pod \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.307603 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-service-ca\") pod \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.307646 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-serving-cert\") pod \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.307677 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-oauth-config\") pod \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.307715 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjqfx\" (UniqueName: \"kubernetes.io/projected/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-kube-api-access-wjqfx\") pod \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\" (UID: \"67d2ef8d-5a0c-48fa-8233-9bb23e794da0\") " Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.308637 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-config" (OuterVolumeSpecName: "console-config") pod "67d2ef8d-5a0c-48fa-8233-9bb23e794da0" (UID: "67d2ef8d-5a0c-48fa-8233-9bb23e794da0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.308653 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-service-ca" (OuterVolumeSpecName: "service-ca") pod "67d2ef8d-5a0c-48fa-8233-9bb23e794da0" (UID: "67d2ef8d-5a0c-48fa-8233-9bb23e794da0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.308663 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "67d2ef8d-5a0c-48fa-8233-9bb23e794da0" (UID: "67d2ef8d-5a0c-48fa-8233-9bb23e794da0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.308850 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "67d2ef8d-5a0c-48fa-8233-9bb23e794da0" (UID: "67d2ef8d-5a0c-48fa-8233-9bb23e794da0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.315093 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "67d2ef8d-5a0c-48fa-8233-9bb23e794da0" (UID: "67d2ef8d-5a0c-48fa-8233-9bb23e794da0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.315159 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-kube-api-access-wjqfx" (OuterVolumeSpecName: "kube-api-access-wjqfx") pod "67d2ef8d-5a0c-48fa-8233-9bb23e794da0" (UID: "67d2ef8d-5a0c-48fa-8233-9bb23e794da0"). InnerVolumeSpecName "kube-api-access-wjqfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.315253 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "67d2ef8d-5a0c-48fa-8233-9bb23e794da0" (UID: "67d2ef8d-5a0c-48fa-8233-9bb23e794da0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.426850 4771 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.427084 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.427239 4771 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.427302 4771 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.427377 4771 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.427434 4771 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.427501 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjqfx\" (UniqueName: \"kubernetes.io/projected/67d2ef8d-5a0c-48fa-8233-9bb23e794da0-kube-api-access-wjqfx\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.843542 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-r664s_67d2ef8d-5a0c-48fa-8233-9bb23e794da0/console/0.log" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.845023 4771 generic.go:334] "Generic (PLEG): container finished" podID="67d2ef8d-5a0c-48fa-8233-9bb23e794da0" containerID="a118fa7267dd547ef56f471f662395be2a9720da4117eb6fef6014523b8ff519" exitCode=2 Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.845213 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r664s" event={"ID":"67d2ef8d-5a0c-48fa-8233-9bb23e794da0","Type":"ContainerDied","Data":"a118fa7267dd547ef56f471f662395be2a9720da4117eb6fef6014523b8ff519"} Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.845271 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r664s" event={"ID":"67d2ef8d-5a0c-48fa-8233-9bb23e794da0","Type":"ContainerDied","Data":"ff4370f822102dba1a096f940eee0b085bf3681ee09b33d72af5987c86a68c36"} Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.845297 4771 scope.go:117] "RemoveContainer" containerID="a118fa7267dd547ef56f471f662395be2a9720da4117eb6fef6014523b8ff519" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.845210 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r664s" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.864212 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-r664s"] Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.866184 4771 scope.go:117] "RemoveContainer" containerID="a118fa7267dd547ef56f471f662395be2a9720da4117eb6fef6014523b8ff519" Oct 02 09:43:49 crc kubenswrapper[4771]: E1002 09:43:49.866577 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a118fa7267dd547ef56f471f662395be2a9720da4117eb6fef6014523b8ff519\": container with ID starting with a118fa7267dd547ef56f471f662395be2a9720da4117eb6fef6014523b8ff519 not found: ID does not exist" containerID="a118fa7267dd547ef56f471f662395be2a9720da4117eb6fef6014523b8ff519" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.866684 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a118fa7267dd547ef56f471f662395be2a9720da4117eb6fef6014523b8ff519"} err="failed to get container status \"a118fa7267dd547ef56f471f662395be2a9720da4117eb6fef6014523b8ff519\": rpc error: code = NotFound desc = could not find container \"a118fa7267dd547ef56f471f662395be2a9720da4117eb6fef6014523b8ff519\": container with ID starting with a118fa7267dd547ef56f471f662395be2a9720da4117eb6fef6014523b8ff519 not found: ID does not exist" Oct 02 09:43:49 crc kubenswrapper[4771]: I1002 09:43:49.867634 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-r664s"] Oct 02 09:43:51 crc kubenswrapper[4771]: I1002 09:43:51.688756 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d2ef8d-5a0c-48fa-8233-9bb23e794da0" path="/var/lib/kubelet/pods/67d2ef8d-5a0c-48fa-8233-9bb23e794da0/volumes" Oct 02 09:43:54 crc kubenswrapper[4771]: I1002 09:43:54.004508 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:43:54 crc kubenswrapper[4771]: I1002 09:43:54.008737 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-6cd7d74865-xhd2p" Oct 02 09:44:14 crc kubenswrapper[4771]: I1002 09:44:14.939412 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:44:14 crc kubenswrapper[4771]: I1002 09:44:14.970810 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:44:15 crc kubenswrapper[4771]: I1002 09:44:15.017961 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.273808 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-78bb69b5d5-cl8tx"] Oct 02 09:44:43 crc kubenswrapper[4771]: E1002 09:44:43.275532 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="675e6316-9501-4ac0-a134-de142b7bfcd4" containerName="registry" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.275567 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="675e6316-9501-4ac0-a134-de142b7bfcd4" containerName="registry" Oct 02 09:44:43 crc kubenswrapper[4771]: E1002 09:44:43.275589 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d2ef8d-5a0c-48fa-8233-9bb23e794da0" containerName="console" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.275595 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d2ef8d-5a0c-48fa-8233-9bb23e794da0" containerName="console" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.275698 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d2ef8d-5a0c-48fa-8233-9bb23e794da0" containerName="console" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.275711 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="675e6316-9501-4ac0-a134-de142b7bfcd4" containerName="registry" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.276170 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.285852 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-78bb69b5d5-cl8tx"] Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.465945 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-oauth-config\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.466106 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-oauth-serving-cert\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.466180 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-config\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.466251 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhtwg\" (UniqueName: \"kubernetes.io/projected/ff5471c7-c311-4bc5-984a-0793a3b53faf-kube-api-access-mhtwg\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.466272 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-serving-cert\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.466566 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-service-ca\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.466615 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-trusted-ca-bundle\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.567780 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-trusted-ca-bundle\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.567883 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-oauth-config\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.567920 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-oauth-serving-cert\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.567947 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-config\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.567972 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhtwg\" (UniqueName: \"kubernetes.io/projected/ff5471c7-c311-4bc5-984a-0793a3b53faf-kube-api-access-mhtwg\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.567994 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-serving-cert\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.568045 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-service-ca\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.568899 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-oauth-serving-cert\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.568952 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-service-ca\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.569150 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-config\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.569160 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-trusted-ca-bundle\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.574648 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-oauth-config\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.574854 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-serving-cert\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.584988 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhtwg\" (UniqueName: \"kubernetes.io/projected/ff5471c7-c311-4bc5-984a-0793a3b53faf-kube-api-access-mhtwg\") pod \"console-78bb69b5d5-cl8tx\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.639615 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:43 crc kubenswrapper[4771]: I1002 09:44:43.849618 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-78bb69b5d5-cl8tx"] Oct 02 09:44:44 crc kubenswrapper[4771]: I1002 09:44:44.163338 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-78bb69b5d5-cl8tx" event={"ID":"ff5471c7-c311-4bc5-984a-0793a3b53faf","Type":"ContainerStarted","Data":"4417aaef2742c4ec4705b18db998c5d23ca7f126c03c2e21bcd198080fa98bbc"} Oct 02 09:44:44 crc kubenswrapper[4771]: I1002 09:44:44.164010 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-78bb69b5d5-cl8tx" event={"ID":"ff5471c7-c311-4bc5-984a-0793a3b53faf","Type":"ContainerStarted","Data":"0a5dae2359324ea25f8e73e8587a23e5c803ef2ac78907ffc2d61be721871a76"} Oct 02 09:44:44 crc kubenswrapper[4771]: I1002 09:44:44.187006 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-78bb69b5d5-cl8tx" podStartSLOduration=1.186985361 podStartE2EDuration="1.186985361s" podCreationTimestamp="2025-10-02 09:44:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:44:44.182864834 +0000 UTC m=+471.830549911" watchObservedRunningTime="2025-10-02 09:44:44.186985361 +0000 UTC m=+471.834670428" Oct 02 09:44:53 crc kubenswrapper[4771]: I1002 09:44:53.640683 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:53 crc kubenswrapper[4771]: I1002 09:44:53.641652 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:53 crc kubenswrapper[4771]: I1002 09:44:53.645964 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:54 crc kubenswrapper[4771]: I1002 09:44:54.219737 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:44:54 crc kubenswrapper[4771]: I1002 09:44:54.276175 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-698d48b8b5-zhfsf"] Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.137461 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz"] Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.139199 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.141072 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.141364 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.147648 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz"] Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.206919 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5548296e-bb64-4fce-8cc0-26f7d7f213e3-config-volume\") pod \"collect-profiles-29323305-cfprz\" (UID: \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.207010 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5548296e-bb64-4fce-8cc0-26f7d7f213e3-secret-volume\") pod \"collect-profiles-29323305-cfprz\" (UID: \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.207051 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smx5w\" (UniqueName: \"kubernetes.io/projected/5548296e-bb64-4fce-8cc0-26f7d7f213e3-kube-api-access-smx5w\") pod \"collect-profiles-29323305-cfprz\" (UID: \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.308297 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5548296e-bb64-4fce-8cc0-26f7d7f213e3-secret-volume\") pod \"collect-profiles-29323305-cfprz\" (UID: \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.308353 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smx5w\" (UniqueName: \"kubernetes.io/projected/5548296e-bb64-4fce-8cc0-26f7d7f213e3-kube-api-access-smx5w\") pod \"collect-profiles-29323305-cfprz\" (UID: \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.308422 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5548296e-bb64-4fce-8cc0-26f7d7f213e3-config-volume\") pod \"collect-profiles-29323305-cfprz\" (UID: \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.309778 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5548296e-bb64-4fce-8cc0-26f7d7f213e3-config-volume\") pod \"collect-profiles-29323305-cfprz\" (UID: \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.315812 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5548296e-bb64-4fce-8cc0-26f7d7f213e3-secret-volume\") pod \"collect-profiles-29323305-cfprz\" (UID: \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.327256 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smx5w\" (UniqueName: \"kubernetes.io/projected/5548296e-bb64-4fce-8cc0-26f7d7f213e3-kube-api-access-smx5w\") pod \"collect-profiles-29323305-cfprz\" (UID: \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.462751 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" Oct 02 09:45:00 crc kubenswrapper[4771]: I1002 09:45:00.730659 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz"] Oct 02 09:45:01 crc kubenswrapper[4771]: I1002 09:45:01.262811 4771 generic.go:334] "Generic (PLEG): container finished" podID="5548296e-bb64-4fce-8cc0-26f7d7f213e3" containerID="c1594021163a791a0349db261831b382673672a437c2d41c727ee75204223ffe" exitCode=0 Oct 02 09:45:01 crc kubenswrapper[4771]: I1002 09:45:01.262879 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" event={"ID":"5548296e-bb64-4fce-8cc0-26f7d7f213e3","Type":"ContainerDied","Data":"c1594021163a791a0349db261831b382673672a437c2d41c727ee75204223ffe"} Oct 02 09:45:01 crc kubenswrapper[4771]: I1002 09:45:01.263301 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" event={"ID":"5548296e-bb64-4fce-8cc0-26f7d7f213e3","Type":"ContainerStarted","Data":"6888e835e225e9fd84ab70d5eaa4c803b12b8b42c170d125d3cb07a3494eeac2"} Oct 02 09:45:02 crc kubenswrapper[4771]: I1002 09:45:02.481446 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" Oct 02 09:45:02 crc kubenswrapper[4771]: I1002 09:45:02.549352 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5548296e-bb64-4fce-8cc0-26f7d7f213e3-config-volume\") pod \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\" (UID: \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\") " Oct 02 09:45:02 crc kubenswrapper[4771]: I1002 09:45:02.549415 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5548296e-bb64-4fce-8cc0-26f7d7f213e3-secret-volume\") pod \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\" (UID: \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\") " Oct 02 09:45:02 crc kubenswrapper[4771]: I1002 09:45:02.549531 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smx5w\" (UniqueName: \"kubernetes.io/projected/5548296e-bb64-4fce-8cc0-26f7d7f213e3-kube-api-access-smx5w\") pod \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\" (UID: \"5548296e-bb64-4fce-8cc0-26f7d7f213e3\") " Oct 02 09:45:02 crc kubenswrapper[4771]: I1002 09:45:02.550775 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5548296e-bb64-4fce-8cc0-26f7d7f213e3-config-volume" (OuterVolumeSpecName: "config-volume") pod "5548296e-bb64-4fce-8cc0-26f7d7f213e3" (UID: "5548296e-bb64-4fce-8cc0-26f7d7f213e3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:02 crc kubenswrapper[4771]: I1002 09:45:02.556262 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5548296e-bb64-4fce-8cc0-26f7d7f213e3-kube-api-access-smx5w" (OuterVolumeSpecName: "kube-api-access-smx5w") pod "5548296e-bb64-4fce-8cc0-26f7d7f213e3" (UID: "5548296e-bb64-4fce-8cc0-26f7d7f213e3"). InnerVolumeSpecName "kube-api-access-smx5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:02 crc kubenswrapper[4771]: I1002 09:45:02.556632 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5548296e-bb64-4fce-8cc0-26f7d7f213e3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5548296e-bb64-4fce-8cc0-26f7d7f213e3" (UID: "5548296e-bb64-4fce-8cc0-26f7d7f213e3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:02 crc kubenswrapper[4771]: I1002 09:45:02.650639 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smx5w\" (UniqueName: \"kubernetes.io/projected/5548296e-bb64-4fce-8cc0-26f7d7f213e3-kube-api-access-smx5w\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:02 crc kubenswrapper[4771]: I1002 09:45:02.650840 4771 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5548296e-bb64-4fce-8cc0-26f7d7f213e3-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:02 crc kubenswrapper[4771]: I1002 09:45:02.650892 4771 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5548296e-bb64-4fce-8cc0-26f7d7f213e3-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:03 crc kubenswrapper[4771]: I1002 09:45:03.277065 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" event={"ID":"5548296e-bb64-4fce-8cc0-26f7d7f213e3","Type":"ContainerDied","Data":"6888e835e225e9fd84ab70d5eaa4c803b12b8b42c170d125d3cb07a3494eeac2"} Oct 02 09:45:03 crc kubenswrapper[4771]: I1002 09:45:03.277109 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6888e835e225e9fd84ab70d5eaa4c803b12b8b42c170d125d3cb07a3494eeac2" Oct 02 09:45:03 crc kubenswrapper[4771]: I1002 09:45:03.277381 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.319916 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-698d48b8b5-zhfsf" podUID="a9f32682-53d4-4987-9402-265115bc84df" containerName="console" containerID="cri-o://616f6c98734dc2ad5642a6567d01fb31f5bbf884f25f7e3b3217a42128f2b22d" gracePeriod=15 Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.645657 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-698d48b8b5-zhfsf_a9f32682-53d4-4987-9402-265115bc84df/console/0.log" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.645730 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.688862 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-trusted-ca-bundle\") pod \"a9f32682-53d4-4987-9402-265115bc84df\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.688962 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt4kn\" (UniqueName: \"kubernetes.io/projected/a9f32682-53d4-4987-9402-265115bc84df-kube-api-access-mt4kn\") pod \"a9f32682-53d4-4987-9402-265115bc84df\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.689035 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9f32682-53d4-4987-9402-265115bc84df-console-serving-cert\") pod \"a9f32682-53d4-4987-9402-265115bc84df\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.689076 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-console-config\") pod \"a9f32682-53d4-4987-9402-265115bc84df\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.689102 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a9f32682-53d4-4987-9402-265115bc84df-console-oauth-config\") pod \"a9f32682-53d4-4987-9402-265115bc84df\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.689152 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-oauth-serving-cert\") pod \"a9f32682-53d4-4987-9402-265115bc84df\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.689201 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-service-ca\") pod \"a9f32682-53d4-4987-9402-265115bc84df\" (UID: \"a9f32682-53d4-4987-9402-265115bc84df\") " Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.690228 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-service-ca" (OuterVolumeSpecName: "service-ca") pod "a9f32682-53d4-4987-9402-265115bc84df" (UID: "a9f32682-53d4-4987-9402-265115bc84df"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.690493 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-console-config" (OuterVolumeSpecName: "console-config") pod "a9f32682-53d4-4987-9402-265115bc84df" (UID: "a9f32682-53d4-4987-9402-265115bc84df"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.690690 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "a9f32682-53d4-4987-9402-265115bc84df" (UID: "a9f32682-53d4-4987-9402-265115bc84df"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.690803 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "a9f32682-53d4-4987-9402-265115bc84df" (UID: "a9f32682-53d4-4987-9402-265115bc84df"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.696467 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9f32682-53d4-4987-9402-265115bc84df-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "a9f32682-53d4-4987-9402-265115bc84df" (UID: "a9f32682-53d4-4987-9402-265115bc84df"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.696540 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9f32682-53d4-4987-9402-265115bc84df-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "a9f32682-53d4-4987-9402-265115bc84df" (UID: "a9f32682-53d4-4987-9402-265115bc84df"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.702245 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9f32682-53d4-4987-9402-265115bc84df-kube-api-access-mt4kn" (OuterVolumeSpecName: "kube-api-access-mt4kn") pod "a9f32682-53d4-4987-9402-265115bc84df" (UID: "a9f32682-53d4-4987-9402-265115bc84df"). InnerVolumeSpecName "kube-api-access-mt4kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.791303 4771 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9f32682-53d4-4987-9402-265115bc84df-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.791555 4771 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.791592 4771 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a9f32682-53d4-4987-9402-265115bc84df-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.791608 4771 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.791617 4771 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.791625 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9f32682-53d4-4987-9402-265115bc84df-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:19 crc kubenswrapper[4771]: I1002 09:45:19.791634 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt4kn\" (UniqueName: \"kubernetes.io/projected/a9f32682-53d4-4987-9402-265115bc84df-kube-api-access-mt4kn\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:20 crc kubenswrapper[4771]: I1002 09:45:20.391251 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-698d48b8b5-zhfsf_a9f32682-53d4-4987-9402-265115bc84df/console/0.log" Oct 02 09:45:20 crc kubenswrapper[4771]: I1002 09:45:20.391311 4771 generic.go:334] "Generic (PLEG): container finished" podID="a9f32682-53d4-4987-9402-265115bc84df" containerID="616f6c98734dc2ad5642a6567d01fb31f5bbf884f25f7e3b3217a42128f2b22d" exitCode=2 Oct 02 09:45:20 crc kubenswrapper[4771]: I1002 09:45:20.391348 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-698d48b8b5-zhfsf" event={"ID":"a9f32682-53d4-4987-9402-265115bc84df","Type":"ContainerDied","Data":"616f6c98734dc2ad5642a6567d01fb31f5bbf884f25f7e3b3217a42128f2b22d"} Oct 02 09:45:20 crc kubenswrapper[4771]: I1002 09:45:20.391381 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-698d48b8b5-zhfsf" event={"ID":"a9f32682-53d4-4987-9402-265115bc84df","Type":"ContainerDied","Data":"9d92fdc8d41d156779a1f61d0f6624c3a6830aba23ebcd1a98cd8fd104fb08b0"} Oct 02 09:45:20 crc kubenswrapper[4771]: I1002 09:45:20.391402 4771 scope.go:117] "RemoveContainer" containerID="616f6c98734dc2ad5642a6567d01fb31f5bbf884f25f7e3b3217a42128f2b22d" Oct 02 09:45:20 crc kubenswrapper[4771]: I1002 09:45:20.391543 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-698d48b8b5-zhfsf" Oct 02 09:45:20 crc kubenswrapper[4771]: I1002 09:45:20.407992 4771 scope.go:117] "RemoveContainer" containerID="616f6c98734dc2ad5642a6567d01fb31f5bbf884f25f7e3b3217a42128f2b22d" Oct 02 09:45:20 crc kubenswrapper[4771]: E1002 09:45:20.408737 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"616f6c98734dc2ad5642a6567d01fb31f5bbf884f25f7e3b3217a42128f2b22d\": container with ID starting with 616f6c98734dc2ad5642a6567d01fb31f5bbf884f25f7e3b3217a42128f2b22d not found: ID does not exist" containerID="616f6c98734dc2ad5642a6567d01fb31f5bbf884f25f7e3b3217a42128f2b22d" Oct 02 09:45:20 crc kubenswrapper[4771]: I1002 09:45:20.408767 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616f6c98734dc2ad5642a6567d01fb31f5bbf884f25f7e3b3217a42128f2b22d"} err="failed to get container status \"616f6c98734dc2ad5642a6567d01fb31f5bbf884f25f7e3b3217a42128f2b22d\": rpc error: code = NotFound desc = could not find container \"616f6c98734dc2ad5642a6567d01fb31f5bbf884f25f7e3b3217a42128f2b22d\": container with ID starting with 616f6c98734dc2ad5642a6567d01fb31f5bbf884f25f7e3b3217a42128f2b22d not found: ID does not exist" Oct 02 09:45:20 crc kubenswrapper[4771]: I1002 09:45:20.422691 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-698d48b8b5-zhfsf"] Oct 02 09:45:20 crc kubenswrapper[4771]: I1002 09:45:20.426565 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-698d48b8b5-zhfsf"] Oct 02 09:45:21 crc kubenswrapper[4771]: I1002 09:45:21.690110 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9f32682-53d4-4987-9402-265115bc84df" path="/var/lib/kubelet/pods/a9f32682-53d4-4987-9402-265115bc84df/volumes" Oct 02 09:45:42 crc kubenswrapper[4771]: I1002 09:45:42.147551 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:45:42 crc kubenswrapper[4771]: I1002 09:45:42.148339 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:46:12 crc kubenswrapper[4771]: I1002 09:46:12.147115 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:46:12 crc kubenswrapper[4771]: I1002 09:46:12.148205 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:46:42 crc kubenswrapper[4771]: I1002 09:46:42.145972 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:46:42 crc kubenswrapper[4771]: I1002 09:46:42.146856 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:46:42 crc kubenswrapper[4771]: I1002 09:46:42.146907 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:46:42 crc kubenswrapper[4771]: I1002 09:46:42.147566 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f2d56a0106e4d97ef1eaacdf9670a1e221241c6eb398fb493028ecc6d5fd6e25"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:46:42 crc kubenswrapper[4771]: I1002 09:46:42.147625 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://f2d56a0106e4d97ef1eaacdf9670a1e221241c6eb398fb493028ecc6d5fd6e25" gracePeriod=600 Oct 02 09:46:42 crc kubenswrapper[4771]: I1002 09:46:42.863332 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="f2d56a0106e4d97ef1eaacdf9670a1e221241c6eb398fb493028ecc6d5fd6e25" exitCode=0 Oct 02 09:46:42 crc kubenswrapper[4771]: I1002 09:46:42.863414 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"f2d56a0106e4d97ef1eaacdf9670a1e221241c6eb398fb493028ecc6d5fd6e25"} Oct 02 09:46:42 crc kubenswrapper[4771]: I1002 09:46:42.863870 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"8f7959ad99e22e4304788ce28b1466fbafe93b57e3295c1b849950830aa10434"} Oct 02 09:46:42 crc kubenswrapper[4771]: I1002 09:46:42.863897 4771 scope.go:117] "RemoveContainer" containerID="78b7a8d589783481aba6ddc2e3b7cc7710dc210299af82522a169351aa5ee352" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.399921 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g"] Oct 02 09:48:07 crc kubenswrapper[4771]: E1002 09:48:07.400732 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9f32682-53d4-4987-9402-265115bc84df" containerName="console" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.400745 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9f32682-53d4-4987-9402-265115bc84df" containerName="console" Oct 02 09:48:07 crc kubenswrapper[4771]: E1002 09:48:07.400772 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5548296e-bb64-4fce-8cc0-26f7d7f213e3" containerName="collect-profiles" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.400778 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="5548296e-bb64-4fce-8cc0-26f7d7f213e3" containerName="collect-profiles" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.400935 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9f32682-53d4-4987-9402-265115bc84df" containerName="console" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.400946 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="5548296e-bb64-4fce-8cc0-26f7d7f213e3" containerName="collect-profiles" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.401760 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.404322 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.412581 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g"] Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.582189 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/238c1703-8f29-422d-8fcf-cb1498b81108-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g\" (UID: \"238c1703-8f29-422d-8fcf-cb1498b81108\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.582267 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvhxf\" (UniqueName: \"kubernetes.io/projected/238c1703-8f29-422d-8fcf-cb1498b81108-kube-api-access-mvhxf\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g\" (UID: \"238c1703-8f29-422d-8fcf-cb1498b81108\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.583120 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/238c1703-8f29-422d-8fcf-cb1498b81108-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g\" (UID: \"238c1703-8f29-422d-8fcf-cb1498b81108\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.684102 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/238c1703-8f29-422d-8fcf-cb1498b81108-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g\" (UID: \"238c1703-8f29-422d-8fcf-cb1498b81108\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.684206 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/238c1703-8f29-422d-8fcf-cb1498b81108-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g\" (UID: \"238c1703-8f29-422d-8fcf-cb1498b81108\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.684255 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvhxf\" (UniqueName: \"kubernetes.io/projected/238c1703-8f29-422d-8fcf-cb1498b81108-kube-api-access-mvhxf\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g\" (UID: \"238c1703-8f29-422d-8fcf-cb1498b81108\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.684748 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/238c1703-8f29-422d-8fcf-cb1498b81108-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g\" (UID: \"238c1703-8f29-422d-8fcf-cb1498b81108\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.684749 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/238c1703-8f29-422d-8fcf-cb1498b81108-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g\" (UID: \"238c1703-8f29-422d-8fcf-cb1498b81108\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.712993 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvhxf\" (UniqueName: \"kubernetes.io/projected/238c1703-8f29-422d-8fcf-cb1498b81108-kube-api-access-mvhxf\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g\" (UID: \"238c1703-8f29-422d-8fcf-cb1498b81108\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" Oct 02 09:48:07 crc kubenswrapper[4771]: I1002 09:48:07.720842 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" Oct 02 09:48:08 crc kubenswrapper[4771]: I1002 09:48:08.222861 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g"] Oct 02 09:48:08 crc kubenswrapper[4771]: I1002 09:48:08.382187 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" event={"ID":"238c1703-8f29-422d-8fcf-cb1498b81108","Type":"ContainerStarted","Data":"9bbe0b36039b56420afb4b63d0f27ec37f458a583e35be4649f8c6a44fc701f6"} Oct 02 09:48:08 crc kubenswrapper[4771]: I1002 09:48:08.382234 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" event={"ID":"238c1703-8f29-422d-8fcf-cb1498b81108","Type":"ContainerStarted","Data":"e4def9991d8ce32ca4cd3c5fcb1beb17ff6c40a80fcc4cd3fc55870e4cda414d"} Oct 02 09:48:09 crc kubenswrapper[4771]: I1002 09:48:09.391642 4771 generic.go:334] "Generic (PLEG): container finished" podID="238c1703-8f29-422d-8fcf-cb1498b81108" containerID="9bbe0b36039b56420afb4b63d0f27ec37f458a583e35be4649f8c6a44fc701f6" exitCode=0 Oct 02 09:48:09 crc kubenswrapper[4771]: I1002 09:48:09.391880 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" event={"ID":"238c1703-8f29-422d-8fcf-cb1498b81108","Type":"ContainerDied","Data":"9bbe0b36039b56420afb4b63d0f27ec37f458a583e35be4649f8c6a44fc701f6"} Oct 02 09:48:09 crc kubenswrapper[4771]: I1002 09:48:09.393769 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:48:11 crc kubenswrapper[4771]: I1002 09:48:11.404783 4771 generic.go:334] "Generic (PLEG): container finished" podID="238c1703-8f29-422d-8fcf-cb1498b81108" containerID="c3422fb345916e6d830aebbbf801165a5af580e17079be1d3bd34fdb2fb1b9d4" exitCode=0 Oct 02 09:48:11 crc kubenswrapper[4771]: I1002 09:48:11.404888 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" event={"ID":"238c1703-8f29-422d-8fcf-cb1498b81108","Type":"ContainerDied","Data":"c3422fb345916e6d830aebbbf801165a5af580e17079be1d3bd34fdb2fb1b9d4"} Oct 02 09:48:12 crc kubenswrapper[4771]: I1002 09:48:12.415988 4771 generic.go:334] "Generic (PLEG): container finished" podID="238c1703-8f29-422d-8fcf-cb1498b81108" containerID="a7cd210e6dbd5abc06cad500cb0090131566b4f51938b8697c3a1ab4f03b0e52" exitCode=0 Oct 02 09:48:12 crc kubenswrapper[4771]: I1002 09:48:12.416094 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" event={"ID":"238c1703-8f29-422d-8fcf-cb1498b81108","Type":"ContainerDied","Data":"a7cd210e6dbd5abc06cad500cb0090131566b4f51938b8697c3a1ab4f03b0e52"} Oct 02 09:48:13 crc kubenswrapper[4771]: I1002 09:48:13.630201 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" Oct 02 09:48:13 crc kubenswrapper[4771]: I1002 09:48:13.804202 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/238c1703-8f29-422d-8fcf-cb1498b81108-util\") pod \"238c1703-8f29-422d-8fcf-cb1498b81108\" (UID: \"238c1703-8f29-422d-8fcf-cb1498b81108\") " Oct 02 09:48:13 crc kubenswrapper[4771]: I1002 09:48:13.804410 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/238c1703-8f29-422d-8fcf-cb1498b81108-bundle\") pod \"238c1703-8f29-422d-8fcf-cb1498b81108\" (UID: \"238c1703-8f29-422d-8fcf-cb1498b81108\") " Oct 02 09:48:13 crc kubenswrapper[4771]: I1002 09:48:13.804448 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvhxf\" (UniqueName: \"kubernetes.io/projected/238c1703-8f29-422d-8fcf-cb1498b81108-kube-api-access-mvhxf\") pod \"238c1703-8f29-422d-8fcf-cb1498b81108\" (UID: \"238c1703-8f29-422d-8fcf-cb1498b81108\") " Oct 02 09:48:13 crc kubenswrapper[4771]: I1002 09:48:13.806410 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/238c1703-8f29-422d-8fcf-cb1498b81108-bundle" (OuterVolumeSpecName: "bundle") pod "238c1703-8f29-422d-8fcf-cb1498b81108" (UID: "238c1703-8f29-422d-8fcf-cb1498b81108"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:13 crc kubenswrapper[4771]: I1002 09:48:13.810359 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/238c1703-8f29-422d-8fcf-cb1498b81108-kube-api-access-mvhxf" (OuterVolumeSpecName: "kube-api-access-mvhxf") pod "238c1703-8f29-422d-8fcf-cb1498b81108" (UID: "238c1703-8f29-422d-8fcf-cb1498b81108"). InnerVolumeSpecName "kube-api-access-mvhxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:13 crc kubenswrapper[4771]: I1002 09:48:13.906098 4771 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/238c1703-8f29-422d-8fcf-cb1498b81108-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:13 crc kubenswrapper[4771]: I1002 09:48:13.906157 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvhxf\" (UniqueName: \"kubernetes.io/projected/238c1703-8f29-422d-8fcf-cb1498b81108-kube-api-access-mvhxf\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:14 crc kubenswrapper[4771]: I1002 09:48:14.086007 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/238c1703-8f29-422d-8fcf-cb1498b81108-util" (OuterVolumeSpecName: "util") pod "238c1703-8f29-422d-8fcf-cb1498b81108" (UID: "238c1703-8f29-422d-8fcf-cb1498b81108"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:14 crc kubenswrapper[4771]: I1002 09:48:14.109271 4771 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/238c1703-8f29-422d-8fcf-cb1498b81108-util\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:14 crc kubenswrapper[4771]: I1002 09:48:14.431340 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" event={"ID":"238c1703-8f29-422d-8fcf-cb1498b81108","Type":"ContainerDied","Data":"e4def9991d8ce32ca4cd3c5fcb1beb17ff6c40a80fcc4cd3fc55870e4cda414d"} Oct 02 09:48:14 crc kubenswrapper[4771]: I1002 09:48:14.431384 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4def9991d8ce32ca4cd3c5fcb1beb17ff6c40a80fcc4cd3fc55870e4cda414d" Oct 02 09:48:14 crc kubenswrapper[4771]: I1002 09:48:14.431667 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g" Oct 02 09:48:18 crc kubenswrapper[4771]: I1002 09:48:18.919318 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4lxxg"] Oct 02 09:48:18 crc kubenswrapper[4771]: I1002 09:48:18.921102 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovn-controller" containerID="cri-o://3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2" gracePeriod=30 Oct 02 09:48:18 crc kubenswrapper[4771]: I1002 09:48:18.921550 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="sbdb" containerID="cri-o://6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa" gracePeriod=30 Oct 02 09:48:18 crc kubenswrapper[4771]: I1002 09:48:18.921675 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392" gracePeriod=30 Oct 02 09:48:18 crc kubenswrapper[4771]: I1002 09:48:18.921702 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="northd" containerID="cri-o://7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85" gracePeriod=30 Oct 02 09:48:18 crc kubenswrapper[4771]: I1002 09:48:18.921726 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovn-acl-logging" containerID="cri-o://86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3" gracePeriod=30 Oct 02 09:48:18 crc kubenswrapper[4771]: I1002 09:48:18.921763 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="nbdb" containerID="cri-o://b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802" gracePeriod=30 Oct 02 09:48:18 crc kubenswrapper[4771]: I1002 09:48:18.921821 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="kube-rbac-proxy-node" containerID="cri-o://82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086" gracePeriod=30 Oct 02 09:48:18 crc kubenswrapper[4771]: I1002 09:48:18.973601 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" containerID="cri-o://43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251" gracePeriod=30 Oct 02 09:48:19 crc kubenswrapper[4771]: E1002 09:48:19.372682 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802 is running failed: container process not found" containerID="b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 02 09:48:19 crc kubenswrapper[4771]: E1002 09:48:19.373007 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802 is running failed: container process not found" containerID="b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 02 09:48:19 crc kubenswrapper[4771]: E1002 09:48:19.373723 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802 is running failed: container process not found" containerID="b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 02 09:48:19 crc kubenswrapper[4771]: E1002 09:48:19.373763 4771 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="nbdb" Oct 02 09:48:19 crc kubenswrapper[4771]: E1002 09:48:19.373864 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 02 09:48:19 crc kubenswrapper[4771]: E1002 09:48:19.374815 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 02 09:48:19 crc kubenswrapper[4771]: E1002 09:48:19.376014 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 02 09:48:19 crc kubenswrapper[4771]: E1002 09:48:19.376063 4771 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="sbdb" Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.461139 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovnkube-controller/3.log" Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.462640 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovn-acl-logging/0.log" Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.462969 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovn-controller/0.log" Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.463390 4771 generic.go:334] "Generic (PLEG): container finished" podID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerID="43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251" exitCode=0 Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.463411 4771 generic.go:334] "Generic (PLEG): container finished" podID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerID="6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa" exitCode=0 Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.463418 4771 generic.go:334] "Generic (PLEG): container finished" podID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerID="b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802" exitCode=0 Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.463426 4771 generic.go:334] "Generic (PLEG): container finished" podID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerID="7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85" exitCode=0 Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.463433 4771 generic.go:334] "Generic (PLEG): container finished" podID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerID="86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3" exitCode=143 Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.463440 4771 generic.go:334] "Generic (PLEG): container finished" podID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerID="3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2" exitCode=143 Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.463482 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251"} Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.463510 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa"} Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.463520 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802"} Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.463529 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85"} Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.463537 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3"} Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.463545 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2"} Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.463559 4771 scope.go:117] "RemoveContainer" containerID="777cad74c30ebd8b7a1585395d702f9a7921e7b6cb303a336d37cb11b46a4eaa" Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.466988 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xh7xv_bb6b65c1-83b7-4b23-abe8-c4255c6e59f8/kube-multus/2.log" Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.467698 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xh7xv_bb6b65c1-83b7-4b23-abe8-c4255c6e59f8/kube-multus/1.log" Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.467726 4771 generic.go:334] "Generic (PLEG): container finished" podID="bb6b65c1-83b7-4b23-abe8-c4255c6e59f8" containerID="c35db692e359d56ccbb79957ddfb37fbf56b3ad4e46093fa55f540da775d2441" exitCode=2 Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.467751 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xh7xv" event={"ID":"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8","Type":"ContainerDied","Data":"c35db692e359d56ccbb79957ddfb37fbf56b3ad4e46093fa55f540da775d2441"} Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.468224 4771 scope.go:117] "RemoveContainer" containerID="c35db692e359d56ccbb79957ddfb37fbf56b3ad4e46093fa55f540da775d2441" Oct 02 09:48:19 crc kubenswrapper[4771]: E1002 09:48:19.468442 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-xh7xv_openshift-multus(bb6b65c1-83b7-4b23-abe8-c4255c6e59f8)\"" pod="openshift-multus/multus-xh7xv" podUID="bb6b65c1-83b7-4b23-abe8-c4255c6e59f8" Oct 02 09:48:19 crc kubenswrapper[4771]: I1002 09:48:19.490234 4771 scope.go:117] "RemoveContainer" containerID="f311d21ae303e154a68174a305b7f431b68a36d9a03fd66eaa55699ade89bcd9" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.131529 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovn-acl-logging/0.log" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.132146 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovn-controller/0.log" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.132576 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284305 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-var-lib-openvswitch\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284371 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovnkube-script-lib\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284409 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-ovn\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284433 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-openvswitch\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284465 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-kubelet\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284483 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-run-ovn-kubernetes\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284511 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-systemd-units\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284533 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-node-log\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284562 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-env-overrides\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284589 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovn-node-metrics-cert\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284627 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-slash\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284647 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-cni-netd\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284669 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovnkube-config\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284689 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-run-netns\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284716 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl2v4\" (UniqueName: \"kubernetes.io/projected/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-kube-api-access-pl2v4\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284754 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-etc-openvswitch\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284786 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-cni-bin\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284824 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-log-socket\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284858 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-systemd\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.284881 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-var-lib-cni-networks-ovn-kubernetes\") pod \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\" (UID: \"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618\") " Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.285225 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.285260 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.285698 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.285731 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.285752 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.285772 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.285792 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.285811 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.285831 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-node-log" (OuterVolumeSpecName: "node-log") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.286052 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.287087 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.287226 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-slash" (OuterVolumeSpecName: "host-slash") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.287305 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.287371 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.287447 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-log-socket" (OuterVolumeSpecName: "log-socket") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.287877 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.287964 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.292405 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.303402 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.303512 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-kube-api-access-pl2v4" (OuterVolumeSpecName: "kube-api-access-pl2v4") pod "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" (UID: "2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618"). InnerVolumeSpecName "kube-api-access-pl2v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.336913 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2drsv"] Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337146 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="kubecfg-setup" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337162 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="kubecfg-setup" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337172 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="northd" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337179 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="northd" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337188 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337194 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337202 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337209 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337223 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337228 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337236 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="238c1703-8f29-422d-8fcf-cb1498b81108" containerName="pull" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337242 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="238c1703-8f29-422d-8fcf-cb1498b81108" containerName="pull" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337251 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337257 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337263 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovn-acl-logging" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337269 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovn-acl-logging" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337279 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="238c1703-8f29-422d-8fcf-cb1498b81108" containerName="util" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337284 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="238c1703-8f29-422d-8fcf-cb1498b81108" containerName="util" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337292 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="kube-rbac-proxy-node" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337297 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="kube-rbac-proxy-node" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337307 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="nbdb" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337312 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="nbdb" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337321 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="238c1703-8f29-422d-8fcf-cb1498b81108" containerName="extract" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337326 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="238c1703-8f29-422d-8fcf-cb1498b81108" containerName="extract" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337334 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovn-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337340 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovn-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337350 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="sbdb" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337359 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="sbdb" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337470 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovn-acl-logging" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337481 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337490 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="kube-rbac-proxy-node" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337501 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337509 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="sbdb" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337519 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337527 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovn-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337536 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="northd" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337544 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337550 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337557 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="238c1703-8f29-422d-8fcf-cb1498b81108" containerName="extract" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337562 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="nbdb" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337657 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337664 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.337672 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337677 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.337790 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerName="ovnkube-controller" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.339653 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.386907 4771 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-slash\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387106 4771 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387182 4771 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387255 4771 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387315 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pl2v4\" (UniqueName: \"kubernetes.io/projected/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-kube-api-access-pl2v4\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387383 4771 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387443 4771 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387492 4771 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-log-socket\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387542 4771 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387590 4771 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387639 4771 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387693 4771 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387741 4771 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387787 4771 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387834 4771 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387882 4771 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387932 4771 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.387984 4771 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-node-log\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.388033 4771 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.388079 4771 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.475161 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xh7xv_bb6b65c1-83b7-4b23-abe8-c4255c6e59f8/kube-multus/2.log" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.479226 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovn-acl-logging/0.log" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.479774 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4lxxg_2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/ovn-controller/0.log" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.480301 4771 generic.go:334] "Generic (PLEG): container finished" podID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerID="f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392" exitCode=0 Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.480382 4771 generic.go:334] "Generic (PLEG): container finished" podID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" containerID="82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086" exitCode=0 Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.480424 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.480375 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392"} Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.480568 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086"} Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.480588 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lxxg" event={"ID":"2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618","Type":"ContainerDied","Data":"0db0a7b33494b001ac2efb0394aa34a8b70b62735c35c50f0e85d8af47548847"} Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.480619 4771 scope.go:117] "RemoveContainer" containerID="43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489224 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-systemd-units\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489284 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-run-netns\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489308 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-run-ovn-kubernetes\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489333 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq659\" (UniqueName: \"kubernetes.io/projected/692afb2e-da73-4735-94e2-df0b178ad1be-kube-api-access-hq659\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489360 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-slash\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489379 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/692afb2e-da73-4735-94e2-df0b178ad1be-ovn-node-metrics-cert\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489398 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-etc-openvswitch\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489418 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-run-systemd\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489439 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/692afb2e-da73-4735-94e2-df0b178ad1be-env-overrides\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489456 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-log-socket\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489485 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/692afb2e-da73-4735-94e2-df0b178ad1be-ovnkube-config\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489513 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-node-log\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489536 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-kubelet\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489556 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489611 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-cni-netd\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489631 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-run-openvswitch\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489652 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-var-lib-openvswitch\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489685 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-cni-bin\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489707 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/692afb2e-da73-4735-94e2-df0b178ad1be-ovnkube-script-lib\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.489727 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-run-ovn\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.496871 4771 scope.go:117] "RemoveContainer" containerID="6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.512886 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4lxxg"] Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.515268 4771 scope.go:117] "RemoveContainer" containerID="b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.517694 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4lxxg"] Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.531969 4771 scope.go:117] "RemoveContainer" containerID="7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.545778 4771 scope.go:117] "RemoveContainer" containerID="f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.558468 4771 scope.go:117] "RemoveContainer" containerID="82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.574297 4771 scope.go:117] "RemoveContainer" containerID="86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591141 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/692afb2e-da73-4735-94e2-df0b178ad1be-ovnkube-config\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591204 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-node-log\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591233 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-kubelet\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591259 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591326 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-run-openvswitch\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591346 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-cni-netd\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591347 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591366 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-var-lib-openvswitch\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591394 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-cni-netd\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591396 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-kubelet\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591438 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-cni-bin\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591451 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-var-lib-openvswitch\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591452 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-run-openvswitch\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591484 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-cni-bin\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591478 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/692afb2e-da73-4735-94e2-df0b178ad1be-ovnkube-script-lib\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591309 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-node-log\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591563 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-run-ovn\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591633 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-run-ovn\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.591688 4771 scope.go:117] "RemoveContainer" containerID="3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592007 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/692afb2e-da73-4735-94e2-df0b178ad1be-ovnkube-config\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592085 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/692afb2e-da73-4735-94e2-df0b178ad1be-ovnkube-script-lib\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592172 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-systemd-units\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592233 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-systemd-units\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592280 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-run-netns\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592304 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-run-ovn-kubernetes\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592347 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-run-netns\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592374 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq659\" (UniqueName: \"kubernetes.io/projected/692afb2e-da73-4735-94e2-df0b178ad1be-kube-api-access-hq659\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592393 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-slash\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592447 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-slash\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592487 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-host-run-ovn-kubernetes\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592411 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-etc-openvswitch\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592524 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/692afb2e-da73-4735-94e2-df0b178ad1be-ovn-node-metrics-cert\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592576 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-etc-openvswitch\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.592783 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-run-systemd\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.593207 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-run-systemd\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.593237 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/692afb2e-da73-4735-94e2-df0b178ad1be-env-overrides\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.593252 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-log-socket\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.593321 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/692afb2e-da73-4735-94e2-df0b178ad1be-log-socket\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.593671 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/692afb2e-da73-4735-94e2-df0b178ad1be-env-overrides\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.615966 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/692afb2e-da73-4735-94e2-df0b178ad1be-ovn-node-metrics-cert\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.616986 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq659\" (UniqueName: \"kubernetes.io/projected/692afb2e-da73-4735-94e2-df0b178ad1be-kube-api-access-hq659\") pod \"ovnkube-node-2drsv\" (UID: \"692afb2e-da73-4735-94e2-df0b178ad1be\") " pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.633453 4771 scope.go:117] "RemoveContainer" containerID="f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.653344 4771 scope.go:117] "RemoveContainer" containerID="43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.653821 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251\": container with ID starting with 43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251 not found: ID does not exist" containerID="43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.653859 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251"} err="failed to get container status \"43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251\": rpc error: code = NotFound desc = could not find container \"43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251\": container with ID starting with 43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.653890 4771 scope.go:117] "RemoveContainer" containerID="6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.654415 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\": container with ID starting with 6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa not found: ID does not exist" containerID="6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.654441 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa"} err="failed to get container status \"6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\": rpc error: code = NotFound desc = could not find container \"6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\": container with ID starting with 6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.654459 4771 scope.go:117] "RemoveContainer" containerID="b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.654799 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\": container with ID starting with b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802 not found: ID does not exist" containerID="b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.654822 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802"} err="failed to get container status \"b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\": rpc error: code = NotFound desc = could not find container \"b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\": container with ID starting with b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.654835 4771 scope.go:117] "RemoveContainer" containerID="7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.655243 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\": container with ID starting with 7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85 not found: ID does not exist" containerID="7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.655272 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85"} err="failed to get container status \"7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\": rpc error: code = NotFound desc = could not find container \"7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\": container with ID starting with 7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.655288 4771 scope.go:117] "RemoveContainer" containerID="f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.655550 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\": container with ID starting with f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392 not found: ID does not exist" containerID="f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.655577 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392"} err="failed to get container status \"f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\": rpc error: code = NotFound desc = could not find container \"f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\": container with ID starting with f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.655594 4771 scope.go:117] "RemoveContainer" containerID="82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.655916 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\": container with ID starting with 82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086 not found: ID does not exist" containerID="82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.655948 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086"} err="failed to get container status \"82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\": rpc error: code = NotFound desc = could not find container \"82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\": container with ID starting with 82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.655970 4771 scope.go:117] "RemoveContainer" containerID="86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.656282 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\": container with ID starting with 86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3 not found: ID does not exist" containerID="86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.656307 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3"} err="failed to get container status \"86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\": rpc error: code = NotFound desc = could not find container \"86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\": container with ID starting with 86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.656320 4771 scope.go:117] "RemoveContainer" containerID="3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.656584 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\": container with ID starting with 3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2 not found: ID does not exist" containerID="3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.656614 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2"} err="failed to get container status \"3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\": rpc error: code = NotFound desc = could not find container \"3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\": container with ID starting with 3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.656629 4771 scope.go:117] "RemoveContainer" containerID="f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68" Oct 02 09:48:20 crc kubenswrapper[4771]: E1002 09:48:20.657021 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\": container with ID starting with f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68 not found: ID does not exist" containerID="f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.657040 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68"} err="failed to get container status \"f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\": rpc error: code = NotFound desc = could not find container \"f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\": container with ID starting with f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.657053 4771 scope.go:117] "RemoveContainer" containerID="43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.657378 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251"} err="failed to get container status \"43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251\": rpc error: code = NotFound desc = could not find container \"43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251\": container with ID starting with 43ad2947fa377e6a4206418757f080adcd267d7cc52f5a827ab723f2d0487251 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.657402 4771 scope.go:117] "RemoveContainer" containerID="6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.657770 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa"} err="failed to get container status \"6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\": rpc error: code = NotFound desc = could not find container \"6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa\": container with ID starting with 6127a0a4323759d8a5a32a38a71ce47e2e31bb82542b53f616bde7028dca75aa not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.657793 4771 scope.go:117] "RemoveContainer" containerID="b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.658231 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802"} err="failed to get container status \"b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\": rpc error: code = NotFound desc = could not find container \"b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802\": container with ID starting with b58f5d81675115c4a77d8436bfadd071e9a823a5399c5ebeebe6a56f4dabe802 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.658253 4771 scope.go:117] "RemoveContainer" containerID="7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.658598 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85"} err="failed to get container status \"7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\": rpc error: code = NotFound desc = could not find container \"7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85\": container with ID starting with 7aa1727c2cab2330ea0b0ac7e38680c4123a7310c89b36d270d7a134225adc85 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.658617 4771 scope.go:117] "RemoveContainer" containerID="f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.658841 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392"} err="failed to get container status \"f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\": rpc error: code = NotFound desc = could not find container \"f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392\": container with ID starting with f2e32e1b8f81c1d5f78ce2f2c5b613dfd62fcb6721b4029b802cda920c1a4392 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.658862 4771 scope.go:117] "RemoveContainer" containerID="82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.659070 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086"} err="failed to get container status \"82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\": rpc error: code = NotFound desc = could not find container \"82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086\": container with ID starting with 82eec2ce2cd96e62328993f7e90f0727fe483e5035d79a580c0a815d3aaec086 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.659086 4771 scope.go:117] "RemoveContainer" containerID="86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.659315 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3"} err="failed to get container status \"86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\": rpc error: code = NotFound desc = could not find container \"86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3\": container with ID starting with 86bbd0342eab7b2d90036bf0da52ed30023cb306ef1ad4c3b9ef5b23d9e476c3 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.659331 4771 scope.go:117] "RemoveContainer" containerID="3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.659499 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2"} err="failed to get container status \"3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\": rpc error: code = NotFound desc = could not find container \"3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2\": container with ID starting with 3ede7569b011a216dd2e135f56d68b2ed0926512c5f441c4d987bdf2618914a2 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.659518 4771 scope.go:117] "RemoveContainer" containerID="f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.659697 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68"} err="failed to get container status \"f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\": rpc error: code = NotFound desc = could not find container \"f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68\": container with ID starting with f4029331dc941189c41148dbe15453ce6ec60355c0529c5daa3a62bbc1aa3f68 not found: ID does not exist" Oct 02 09:48:20 crc kubenswrapper[4771]: I1002 09:48:20.660015 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:21 crc kubenswrapper[4771]: I1002 09:48:21.487750 4771 generic.go:334] "Generic (PLEG): container finished" podID="692afb2e-da73-4735-94e2-df0b178ad1be" containerID="a5f8182cd8849c070162bff1c69ad1a8c6fa2af4ac05b98b3bed835a931df2de" exitCode=0 Oct 02 09:48:21 crc kubenswrapper[4771]: I1002 09:48:21.487836 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" event={"ID":"692afb2e-da73-4735-94e2-df0b178ad1be","Type":"ContainerDied","Data":"a5f8182cd8849c070162bff1c69ad1a8c6fa2af4ac05b98b3bed835a931df2de"} Oct 02 09:48:21 crc kubenswrapper[4771]: I1002 09:48:21.488244 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" event={"ID":"692afb2e-da73-4735-94e2-df0b178ad1be","Type":"ContainerStarted","Data":"97bcd918e608b212d0572ccfd263be739042c96acfdaa545775cfb0cd5de8e9d"} Oct 02 09:48:21 crc kubenswrapper[4771]: I1002 09:48:21.698700 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618" path="/var/lib/kubelet/pods/2c5d0dc7-d6bf-4c34-aba5-a14dbdbde618/volumes" Oct 02 09:48:22 crc kubenswrapper[4771]: I1002 09:48:22.496563 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" event={"ID":"692afb2e-da73-4735-94e2-df0b178ad1be","Type":"ContainerStarted","Data":"54669ad4b22e13fdd0dad624bc23c0826a454896ccfd7f8b82a3d1801b757261"} Oct 02 09:48:22 crc kubenswrapper[4771]: I1002 09:48:22.496902 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" event={"ID":"692afb2e-da73-4735-94e2-df0b178ad1be","Type":"ContainerStarted","Data":"d6e62e59a5330de56d3bed83e149599c73084915bc175e57c307b87c85ac56a5"} Oct 02 09:48:22 crc kubenswrapper[4771]: I1002 09:48:22.496916 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" event={"ID":"692afb2e-da73-4735-94e2-df0b178ad1be","Type":"ContainerStarted","Data":"c0f5aa9775cd389ef744dfafaee6d905a9eb3eef7a09d6d3842cdc90cccff642"} Oct 02 09:48:22 crc kubenswrapper[4771]: I1002 09:48:22.496927 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" event={"ID":"692afb2e-da73-4735-94e2-df0b178ad1be","Type":"ContainerStarted","Data":"b9da595a17cedc04e03512387a8e3aa56c39c6a77b72b2c83ab76baaebc6ed29"} Oct 02 09:48:22 crc kubenswrapper[4771]: I1002 09:48:22.496939 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" event={"ID":"692afb2e-da73-4735-94e2-df0b178ad1be","Type":"ContainerStarted","Data":"2f5bb6f538e1379fd154807451e728e5b1399c25203dcb94046786641989f4e8"} Oct 02 09:48:23 crc kubenswrapper[4771]: I1002 09:48:23.507187 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" event={"ID":"692afb2e-da73-4735-94e2-df0b178ad1be","Type":"ContainerStarted","Data":"2417f30bc4570b1494f21cba3f26d96853dbd56792edd4e5638f78ceefd0a94d"} Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.603151 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps"] Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.604161 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.606625 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.606678 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-nvfnh" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.609498 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.650422 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhsd8\" (UniqueName: \"kubernetes.io/projected/9c0d8940-6aa3-461a-9b29-53277779bb47-kube-api-access-bhsd8\") pod \"obo-prometheus-operator-7c8cf85677-tbgps\" (UID: \"9c0d8940-6aa3-461a-9b29-53277779bb47\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.736733 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn"] Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.737826 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.742822 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-h2mqc" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.743249 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.751501 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8fd0bf98-7976-4e81-8afa-76b37aff3944-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn\" (UID: \"8fd0bf98-7976-4e81-8afa-76b37aff3944\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.751566 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhsd8\" (UniqueName: \"kubernetes.io/projected/9c0d8940-6aa3-461a-9b29-53277779bb47-kube-api-access-bhsd8\") pod \"obo-prometheus-operator-7c8cf85677-tbgps\" (UID: \"9c0d8940-6aa3-461a-9b29-53277779bb47\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.751631 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8fd0bf98-7976-4e81-8afa-76b37aff3944-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn\" (UID: \"8fd0bf98-7976-4e81-8afa-76b37aff3944\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.768614 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2"] Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.769965 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.779444 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhsd8\" (UniqueName: \"kubernetes.io/projected/9c0d8940-6aa3-461a-9b29-53277779bb47-kube-api-access-bhsd8\") pod \"obo-prometheus-operator-7c8cf85677-tbgps\" (UID: \"9c0d8940-6aa3-461a-9b29-53277779bb47\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.852454 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6214e26e-87c0-4bfb-bb9f-889636a4a9fa-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2\" (UID: \"6214e26e-87c0-4bfb-bb9f-889636a4a9fa\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.852507 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8fd0bf98-7976-4e81-8afa-76b37aff3944-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn\" (UID: \"8fd0bf98-7976-4e81-8afa-76b37aff3944\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.852533 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6214e26e-87c0-4bfb-bb9f-889636a4a9fa-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2\" (UID: \"6214e26e-87c0-4bfb-bb9f-889636a4a9fa\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.852781 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8fd0bf98-7976-4e81-8afa-76b37aff3944-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn\" (UID: \"8fd0bf98-7976-4e81-8afa-76b37aff3944\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.857297 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8fd0bf98-7976-4e81-8afa-76b37aff3944-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn\" (UID: \"8fd0bf98-7976-4e81-8afa-76b37aff3944\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.857492 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8fd0bf98-7976-4e81-8afa-76b37aff3944-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn\" (UID: \"8fd0bf98-7976-4e81-8afa-76b37aff3944\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.923350 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.942373 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-wwgrl"] Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.943318 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.948587 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-2gtws" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.948874 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.954382 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wghr\" (UniqueName: \"kubernetes.io/projected/e792e389-a48d-47fa-8c41-9767cb669865-kube-api-access-5wghr\") pod \"observability-operator-cc5f78dfc-wwgrl\" (UID: \"e792e389-a48d-47fa-8c41-9767cb669865\") " pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.954463 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/e792e389-a48d-47fa-8c41-9767cb669865-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-wwgrl\" (UID: \"e792e389-a48d-47fa-8c41-9767cb669865\") " pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.954520 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6214e26e-87c0-4bfb-bb9f-889636a4a9fa-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2\" (UID: \"6214e26e-87c0-4bfb-bb9f-889636a4a9fa\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.954565 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6214e26e-87c0-4bfb-bb9f-889636a4a9fa-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2\" (UID: \"6214e26e-87c0-4bfb-bb9f-889636a4a9fa\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.963395 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6214e26e-87c0-4bfb-bb9f-889636a4a9fa-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2\" (UID: \"6214e26e-87c0-4bfb-bb9f-889636a4a9fa\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:24 crc kubenswrapper[4771]: I1002 09:48:24.987340 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6214e26e-87c0-4bfb-bb9f-889636a4a9fa-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2\" (UID: \"6214e26e-87c0-4bfb-bb9f-889636a4a9fa\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:24 crc kubenswrapper[4771]: E1002 09:48:24.997354 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators_9c0d8940-6aa3-461a-9b29-53277779bb47_0(21d7c90d35d306b2cf0cd9261c25bbd2c13eb98cf332a8cb1aa63bf410dbfdd6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:24 crc kubenswrapper[4771]: E1002 09:48:24.997437 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators_9c0d8940-6aa3-461a-9b29-53277779bb47_0(21d7c90d35d306b2cf0cd9261c25bbd2c13eb98cf332a8cb1aa63bf410dbfdd6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:24 crc kubenswrapper[4771]: E1002 09:48:24.997461 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators_9c0d8940-6aa3-461a-9b29-53277779bb47_0(21d7c90d35d306b2cf0cd9261c25bbd2c13eb98cf332a8cb1aa63bf410dbfdd6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:24 crc kubenswrapper[4771]: E1002 09:48:24.997530 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators(9c0d8940-6aa3-461a-9b29-53277779bb47)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators(9c0d8940-6aa3-461a-9b29-53277779bb47)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators_9c0d8940-6aa3-461a-9b29-53277779bb47_0(21d7c90d35d306b2cf0cd9261c25bbd2c13eb98cf332a8cb1aa63bf410dbfdd6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" podUID="9c0d8940-6aa3-461a-9b29-53277779bb47" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.050930 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-2ccmn"] Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.051692 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.055361 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpd9n\" (UniqueName: \"kubernetes.io/projected/fa786166-e5d7-43e7-8803-98135ffc2871-kube-api-access-hpd9n\") pod \"perses-operator-54bc95c9fb-2ccmn\" (UID: \"fa786166-e5d7-43e7-8803-98135ffc2871\") " pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.055384 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-5jl96" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.055433 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/e792e389-a48d-47fa-8c41-9767cb669865-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-wwgrl\" (UID: \"e792e389-a48d-47fa-8c41-9767cb669865\") " pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.055466 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/fa786166-e5d7-43e7-8803-98135ffc2871-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-2ccmn\" (UID: \"fa786166-e5d7-43e7-8803-98135ffc2871\") " pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.055523 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wghr\" (UniqueName: \"kubernetes.io/projected/e792e389-a48d-47fa-8c41-9767cb669865-kube-api-access-5wghr\") pod \"observability-operator-cc5f78dfc-wwgrl\" (UID: \"e792e389-a48d-47fa-8c41-9767cb669865\") " pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.058981 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/e792e389-a48d-47fa-8c41-9767cb669865-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-wwgrl\" (UID: \"e792e389-a48d-47fa-8c41-9767cb669865\") " pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.079643 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.084925 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wghr\" (UniqueName: \"kubernetes.io/projected/e792e389-a48d-47fa-8c41-9767cb669865-kube-api-access-5wghr\") pod \"observability-operator-cc5f78dfc-wwgrl\" (UID: \"e792e389-a48d-47fa-8c41-9767cb669865\") " pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.107867 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators_8fd0bf98-7976-4e81-8afa-76b37aff3944_0(4127f1f55df8550a49c01a90cf479432a347fbd65a76f0495456dee56fa726b1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.107934 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators_8fd0bf98-7976-4e81-8afa-76b37aff3944_0(4127f1f55df8550a49c01a90cf479432a347fbd65a76f0495456dee56fa726b1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.107963 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators_8fd0bf98-7976-4e81-8afa-76b37aff3944_0(4127f1f55df8550a49c01a90cf479432a347fbd65a76f0495456dee56fa726b1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.108011 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators(8fd0bf98-7976-4e81-8afa-76b37aff3944)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators(8fd0bf98-7976-4e81-8afa-76b37aff3944)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators_8fd0bf98-7976-4e81-8afa-76b37aff3944_0(4127f1f55df8550a49c01a90cf479432a347fbd65a76f0495456dee56fa726b1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" podUID="8fd0bf98-7976-4e81-8afa-76b37aff3944" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.145087 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.156403 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpd9n\" (UniqueName: \"kubernetes.io/projected/fa786166-e5d7-43e7-8803-98135ffc2871-kube-api-access-hpd9n\") pod \"perses-operator-54bc95c9fb-2ccmn\" (UID: \"fa786166-e5d7-43e7-8803-98135ffc2871\") " pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.156465 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/fa786166-e5d7-43e7-8803-98135ffc2871-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-2ccmn\" (UID: \"fa786166-e5d7-43e7-8803-98135ffc2871\") " pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.157343 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/fa786166-e5d7-43e7-8803-98135ffc2871-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-2ccmn\" (UID: \"fa786166-e5d7-43e7-8803-98135ffc2871\") " pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.164911 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators_6214e26e-87c0-4bfb-bb9f-889636a4a9fa_0(9e9131f4656941843b24c4c8d8e3cc89bfa132dbcd890b2e15e1588961c80da9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.164971 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators_6214e26e-87c0-4bfb-bb9f-889636a4a9fa_0(9e9131f4656941843b24c4c8d8e3cc89bfa132dbcd890b2e15e1588961c80da9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.164993 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators_6214e26e-87c0-4bfb-bb9f-889636a4a9fa_0(9e9131f4656941843b24c4c8d8e3cc89bfa132dbcd890b2e15e1588961c80da9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.165043 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators(6214e26e-87c0-4bfb-bb9f-889636a4a9fa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators(6214e26e-87c0-4bfb-bb9f-889636a4a9fa)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators_6214e26e-87c0-4bfb-bb9f-889636a4a9fa_0(9e9131f4656941843b24c4c8d8e3cc89bfa132dbcd890b2e15e1588961c80da9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" podUID="6214e26e-87c0-4bfb-bb9f-889636a4a9fa" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.174280 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpd9n\" (UniqueName: \"kubernetes.io/projected/fa786166-e5d7-43e7-8803-98135ffc2871-kube-api-access-hpd9n\") pod \"perses-operator-54bc95c9fb-2ccmn\" (UID: \"fa786166-e5d7-43e7-8803-98135ffc2871\") " pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.304169 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.323564 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-cc5f78dfc-wwgrl_openshift-operators_e792e389-a48d-47fa-8c41-9767cb669865_0(12eb007dabeced09299c1ac2f73323a9f930928be25ed681b0904197d0392f13): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.323674 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-cc5f78dfc-wwgrl_openshift-operators_e792e389-a48d-47fa-8c41-9767cb669865_0(12eb007dabeced09299c1ac2f73323a9f930928be25ed681b0904197d0392f13): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.323702 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-cc5f78dfc-wwgrl_openshift-operators_e792e389-a48d-47fa-8c41-9767cb669865_0(12eb007dabeced09299c1ac2f73323a9f930928be25ed681b0904197d0392f13): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.323759 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-cc5f78dfc-wwgrl_openshift-operators(e792e389-a48d-47fa-8c41-9767cb669865)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-cc5f78dfc-wwgrl_openshift-operators(e792e389-a48d-47fa-8c41-9767cb669865)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-cc5f78dfc-wwgrl_openshift-operators_e792e389-a48d-47fa-8c41-9767cb669865_0(12eb007dabeced09299c1ac2f73323a9f930928be25ed681b0904197d0392f13): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" podUID="e792e389-a48d-47fa-8c41-9767cb669865" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.368387 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.390438 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-2ccmn_openshift-operators_fa786166-e5d7-43e7-8803-98135ffc2871_0(981d26e1dffe42bbdc04bc6fe689060282b32a8f80d2009e4fbc6f247f6bd3eb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.390534 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-2ccmn_openshift-operators_fa786166-e5d7-43e7-8803-98135ffc2871_0(981d26e1dffe42bbdc04bc6fe689060282b32a8f80d2009e4fbc6f247f6bd3eb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.390562 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-2ccmn_openshift-operators_fa786166-e5d7-43e7-8803-98135ffc2871_0(981d26e1dffe42bbdc04bc6fe689060282b32a8f80d2009e4fbc6f247f6bd3eb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:25 crc kubenswrapper[4771]: E1002 09:48:25.390620 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-54bc95c9fb-2ccmn_openshift-operators(fa786166-e5d7-43e7-8803-98135ffc2871)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-54bc95c9fb-2ccmn_openshift-operators(fa786166-e5d7-43e7-8803-98135ffc2871)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-2ccmn_openshift-operators_fa786166-e5d7-43e7-8803-98135ffc2871_0(981d26e1dffe42bbdc04bc6fe689060282b32a8f80d2009e4fbc6f247f6bd3eb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" podUID="fa786166-e5d7-43e7-8803-98135ffc2871" Oct 02 09:48:25 crc kubenswrapper[4771]: I1002 09:48:25.528398 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" event={"ID":"692afb2e-da73-4735-94e2-df0b178ad1be","Type":"ContainerStarted","Data":"87f2bfcb9b64d70844f88950d23f3930c144bd06837ec2856fd32eede27490e9"} Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.543337 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" event={"ID":"692afb2e-da73-4735-94e2-df0b178ad1be","Type":"ContainerStarted","Data":"3972d46ed90a7b9454a2fe54ef86c60803ac17497099a92820da982e5bf16bf5"} Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.543651 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.543741 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.571758 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.576484 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" podStartSLOduration=7.576462299 podStartE2EDuration="7.576462299s" podCreationTimestamp="2025-10-02 09:48:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:48:27.572007884 +0000 UTC m=+695.219692961" watchObservedRunningTime="2025-10-02 09:48:27.576462299 +0000 UTC m=+695.224147366" Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.955812 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps"] Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.956288 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.956807 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.988898 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2"] Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.989035 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.989549 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.998195 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-wwgrl"] Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.998373 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:27 crc kubenswrapper[4771]: I1002 09:48:27.998907 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:28 crc kubenswrapper[4771]: I1002 09:48:28.011867 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-2ccmn"] Oct 02 09:48:28 crc kubenswrapper[4771]: I1002 09:48:28.011927 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn"] Oct 02 09:48:28 crc kubenswrapper[4771]: I1002 09:48:28.011999 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:28 crc kubenswrapper[4771]: I1002 09:48:28.013215 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:28 crc kubenswrapper[4771]: I1002 09:48:28.013733 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:28 crc kubenswrapper[4771]: I1002 09:48:28.021750 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.055290 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators_9c0d8940-6aa3-461a-9b29-53277779bb47_0(20c3dbe8b9d95439523036f9e6127eb6f396fc728adaa4964079a6f1d5e43e49): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.055368 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators_9c0d8940-6aa3-461a-9b29-53277779bb47_0(20c3dbe8b9d95439523036f9e6127eb6f396fc728adaa4964079a6f1d5e43e49): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.055398 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators_9c0d8940-6aa3-461a-9b29-53277779bb47_0(20c3dbe8b9d95439523036f9e6127eb6f396fc728adaa4964079a6f1d5e43e49): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.055457 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators(9c0d8940-6aa3-461a-9b29-53277779bb47)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators(9c0d8940-6aa3-461a-9b29-53277779bb47)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators_9c0d8940-6aa3-461a-9b29-53277779bb47_0(20c3dbe8b9d95439523036f9e6127eb6f396fc728adaa4964079a6f1d5e43e49): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" podUID="9c0d8940-6aa3-461a-9b29-53277779bb47" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.135496 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators_6214e26e-87c0-4bfb-bb9f-889636a4a9fa_0(e7ed146c14087d9b6c6e1800e49fd0cfd88adc98388a50673f843948be80315e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.135575 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators_6214e26e-87c0-4bfb-bb9f-889636a4a9fa_0(e7ed146c14087d9b6c6e1800e49fd0cfd88adc98388a50673f843948be80315e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.135606 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators_6214e26e-87c0-4bfb-bb9f-889636a4a9fa_0(e7ed146c14087d9b6c6e1800e49fd0cfd88adc98388a50673f843948be80315e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.135670 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators(6214e26e-87c0-4bfb-bb9f-889636a4a9fa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators(6214e26e-87c0-4bfb-bb9f-889636a4a9fa)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators_6214e26e-87c0-4bfb-bb9f-889636a4a9fa_0(e7ed146c14087d9b6c6e1800e49fd0cfd88adc98388a50673f843948be80315e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" podUID="6214e26e-87c0-4bfb-bb9f-889636a4a9fa" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.141418 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-cc5f78dfc-wwgrl_openshift-operators_e792e389-a48d-47fa-8c41-9767cb669865_0(7dfb028f5584fb7d17b0b6593f059561ef1906a5bf70c70f55c99446f88367d8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.141474 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-cc5f78dfc-wwgrl_openshift-operators_e792e389-a48d-47fa-8c41-9767cb669865_0(7dfb028f5584fb7d17b0b6593f059561ef1906a5bf70c70f55c99446f88367d8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.141501 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-cc5f78dfc-wwgrl_openshift-operators_e792e389-a48d-47fa-8c41-9767cb669865_0(7dfb028f5584fb7d17b0b6593f059561ef1906a5bf70c70f55c99446f88367d8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.141542 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-cc5f78dfc-wwgrl_openshift-operators(e792e389-a48d-47fa-8c41-9767cb669865)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-cc5f78dfc-wwgrl_openshift-operators(e792e389-a48d-47fa-8c41-9767cb669865)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-cc5f78dfc-wwgrl_openshift-operators_e792e389-a48d-47fa-8c41-9767cb669865_0(7dfb028f5584fb7d17b0b6593f059561ef1906a5bf70c70f55c99446f88367d8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" podUID="e792e389-a48d-47fa-8c41-9767cb669865" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.161595 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators_8fd0bf98-7976-4e81-8afa-76b37aff3944_0(cb4e2e44e6ac0be2e28ea6f342a1e464db2798c64c6ba17cf86c6419e6f47ba6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.161667 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators_8fd0bf98-7976-4e81-8afa-76b37aff3944_0(cb4e2e44e6ac0be2e28ea6f342a1e464db2798c64c6ba17cf86c6419e6f47ba6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.161693 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators_8fd0bf98-7976-4e81-8afa-76b37aff3944_0(cb4e2e44e6ac0be2e28ea6f342a1e464db2798c64c6ba17cf86c6419e6f47ba6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.161738 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators(8fd0bf98-7976-4e81-8afa-76b37aff3944)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators(8fd0bf98-7976-4e81-8afa-76b37aff3944)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators_8fd0bf98-7976-4e81-8afa-76b37aff3944_0(cb4e2e44e6ac0be2e28ea6f342a1e464db2798c64c6ba17cf86c6419e6f47ba6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" podUID="8fd0bf98-7976-4e81-8afa-76b37aff3944" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.165847 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-2ccmn_openshift-operators_fa786166-e5d7-43e7-8803-98135ffc2871_0(5608ccccd68ec7bd88c1f8fdfc268597d89615946123ec7cad4f29b7a88e2bc2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.165903 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-2ccmn_openshift-operators_fa786166-e5d7-43e7-8803-98135ffc2871_0(5608ccccd68ec7bd88c1f8fdfc268597d89615946123ec7cad4f29b7a88e2bc2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.165926 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-2ccmn_openshift-operators_fa786166-e5d7-43e7-8803-98135ffc2871_0(5608ccccd68ec7bd88c1f8fdfc268597d89615946123ec7cad4f29b7a88e2bc2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:28 crc kubenswrapper[4771]: E1002 09:48:28.165994 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-54bc95c9fb-2ccmn_openshift-operators(fa786166-e5d7-43e7-8803-98135ffc2871)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-54bc95c9fb-2ccmn_openshift-operators(fa786166-e5d7-43e7-8803-98135ffc2871)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-2ccmn_openshift-operators_fa786166-e5d7-43e7-8803-98135ffc2871_0(5608ccccd68ec7bd88c1f8fdfc268597d89615946123ec7cad4f29b7a88e2bc2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" podUID="fa786166-e5d7-43e7-8803-98135ffc2871" Oct 02 09:48:28 crc kubenswrapper[4771]: I1002 09:48:28.548959 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:28 crc kubenswrapper[4771]: I1002 09:48:28.578397 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:31 crc kubenswrapper[4771]: I1002 09:48:31.681466 4771 scope.go:117] "RemoveContainer" containerID="c35db692e359d56ccbb79957ddfb37fbf56b3ad4e46093fa55f540da775d2441" Oct 02 09:48:31 crc kubenswrapper[4771]: E1002 09:48:31.682195 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-xh7xv_openshift-multus(bb6b65c1-83b7-4b23-abe8-c4255c6e59f8)\"" pod="openshift-multus/multus-xh7xv" podUID="bb6b65c1-83b7-4b23-abe8-c4255c6e59f8" Oct 02 09:48:40 crc kubenswrapper[4771]: I1002 09:48:40.680768 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:40 crc kubenswrapper[4771]: I1002 09:48:40.681863 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:40 crc kubenswrapper[4771]: E1002 09:48:40.724773 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators_9c0d8940-6aa3-461a-9b29-53277779bb47_0(fb157d4e323b2898736c8f44e526045926e360836d6e2624e1a16259786c65e7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:40 crc kubenswrapper[4771]: E1002 09:48:40.724973 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators_9c0d8940-6aa3-461a-9b29-53277779bb47_0(fb157d4e323b2898736c8f44e526045926e360836d6e2624e1a16259786c65e7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:40 crc kubenswrapper[4771]: E1002 09:48:40.725211 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators_9c0d8940-6aa3-461a-9b29-53277779bb47_0(fb157d4e323b2898736c8f44e526045926e360836d6e2624e1a16259786c65e7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:40 crc kubenswrapper[4771]: E1002 09:48:40.725376 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators(9c0d8940-6aa3-461a-9b29-53277779bb47)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators(9c0d8940-6aa3-461a-9b29-53277779bb47)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-tbgps_openshift-operators_9c0d8940-6aa3-461a-9b29-53277779bb47_0(fb157d4e323b2898736c8f44e526045926e360836d6e2624e1a16259786c65e7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" podUID="9c0d8940-6aa3-461a-9b29-53277779bb47" Oct 02 09:48:41 crc kubenswrapper[4771]: I1002 09:48:41.680907 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:41 crc kubenswrapper[4771]: I1002 09:48:41.681456 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:41 crc kubenswrapper[4771]: E1002 09:48:41.748085 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-cc5f78dfc-wwgrl_openshift-operators_e792e389-a48d-47fa-8c41-9767cb669865_0(d3d5fea8046eb5d691cec90e8e7f0abec7d65ac2b42db1221756e2d17257d383): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:41 crc kubenswrapper[4771]: E1002 09:48:41.748464 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-cc5f78dfc-wwgrl_openshift-operators_e792e389-a48d-47fa-8c41-9767cb669865_0(d3d5fea8046eb5d691cec90e8e7f0abec7d65ac2b42db1221756e2d17257d383): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:41 crc kubenswrapper[4771]: E1002 09:48:41.748489 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-cc5f78dfc-wwgrl_openshift-operators_e792e389-a48d-47fa-8c41-9767cb669865_0(d3d5fea8046eb5d691cec90e8e7f0abec7d65ac2b42db1221756e2d17257d383): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:41 crc kubenswrapper[4771]: E1002 09:48:41.748540 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-cc5f78dfc-wwgrl_openshift-operators(e792e389-a48d-47fa-8c41-9767cb669865)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-cc5f78dfc-wwgrl_openshift-operators(e792e389-a48d-47fa-8c41-9767cb669865)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-cc5f78dfc-wwgrl_openshift-operators_e792e389-a48d-47fa-8c41-9767cb669865_0(d3d5fea8046eb5d691cec90e8e7f0abec7d65ac2b42db1221756e2d17257d383): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" podUID="e792e389-a48d-47fa-8c41-9767cb669865" Oct 02 09:48:42 crc kubenswrapper[4771]: I1002 09:48:42.146677 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:48:42 crc kubenswrapper[4771]: I1002 09:48:42.146745 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:48:42 crc kubenswrapper[4771]: I1002 09:48:42.680720 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:42 crc kubenswrapper[4771]: I1002 09:48:42.680971 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:42 crc kubenswrapper[4771]: I1002 09:48:42.681073 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:42 crc kubenswrapper[4771]: I1002 09:48:42.681795 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:42 crc kubenswrapper[4771]: I1002 09:48:42.681830 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:42 crc kubenswrapper[4771]: I1002 09:48:42.681832 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:42 crc kubenswrapper[4771]: E1002 09:48:42.724344 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-2ccmn_openshift-operators_fa786166-e5d7-43e7-8803-98135ffc2871_0(abae3e668dec143839006c008ea7d6a3099bfe4cc481fec976817d675b24d70a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:42 crc kubenswrapper[4771]: E1002 09:48:42.724413 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-2ccmn_openshift-operators_fa786166-e5d7-43e7-8803-98135ffc2871_0(abae3e668dec143839006c008ea7d6a3099bfe4cc481fec976817d675b24d70a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:42 crc kubenswrapper[4771]: E1002 09:48:42.724439 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-2ccmn_openshift-operators_fa786166-e5d7-43e7-8803-98135ffc2871_0(abae3e668dec143839006c008ea7d6a3099bfe4cc481fec976817d675b24d70a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:42 crc kubenswrapper[4771]: E1002 09:48:42.724500 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-54bc95c9fb-2ccmn_openshift-operators(fa786166-e5d7-43e7-8803-98135ffc2871)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-54bc95c9fb-2ccmn_openshift-operators(fa786166-e5d7-43e7-8803-98135ffc2871)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-2ccmn_openshift-operators_fa786166-e5d7-43e7-8803-98135ffc2871_0(abae3e668dec143839006c008ea7d6a3099bfe4cc481fec976817d675b24d70a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" podUID="fa786166-e5d7-43e7-8803-98135ffc2871" Oct 02 09:48:42 crc kubenswrapper[4771]: E1002 09:48:42.736768 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators_8fd0bf98-7976-4e81-8afa-76b37aff3944_0(095223c548b8548ff2214ac7637bc4c16379a16c115ec0a29f84bb31f90164ac): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:42 crc kubenswrapper[4771]: E1002 09:48:42.736837 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators_8fd0bf98-7976-4e81-8afa-76b37aff3944_0(095223c548b8548ff2214ac7637bc4c16379a16c115ec0a29f84bb31f90164ac): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:42 crc kubenswrapper[4771]: E1002 09:48:42.736864 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators_8fd0bf98-7976-4e81-8afa-76b37aff3944_0(095223c548b8548ff2214ac7637bc4c16379a16c115ec0a29f84bb31f90164ac): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:42 crc kubenswrapper[4771]: E1002 09:48:42.736930 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators(8fd0bf98-7976-4e81-8afa-76b37aff3944)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators(8fd0bf98-7976-4e81-8afa-76b37aff3944)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_openshift-operators_8fd0bf98-7976-4e81-8afa-76b37aff3944_0(095223c548b8548ff2214ac7637bc4c16379a16c115ec0a29f84bb31f90164ac): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" podUID="8fd0bf98-7976-4e81-8afa-76b37aff3944" Oct 02 09:48:42 crc kubenswrapper[4771]: E1002 09:48:42.743864 4771 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators_6214e26e-87c0-4bfb-bb9f-889636a4a9fa_0(f7f195a2b82ae72b72bde6e1a1e8b1393a4f937fb38b6e9dd33b352065fd9a52): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:48:42 crc kubenswrapper[4771]: E1002 09:48:42.744010 4771 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators_6214e26e-87c0-4bfb-bb9f-889636a4a9fa_0(f7f195a2b82ae72b72bde6e1a1e8b1393a4f937fb38b6e9dd33b352065fd9a52): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:42 crc kubenswrapper[4771]: E1002 09:48:42.744085 4771 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators_6214e26e-87c0-4bfb-bb9f-889636a4a9fa_0(f7f195a2b82ae72b72bde6e1a1e8b1393a4f937fb38b6e9dd33b352065fd9a52): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:42 crc kubenswrapper[4771]: E1002 09:48:42.744252 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators(6214e26e-87c0-4bfb-bb9f-889636a4a9fa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators(6214e26e-87c0-4bfb-bb9f-889636a4a9fa)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_openshift-operators_6214e26e-87c0-4bfb-bb9f-889636a4a9fa_0(f7f195a2b82ae72b72bde6e1a1e8b1393a4f937fb38b6e9dd33b352065fd9a52): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" podUID="6214e26e-87c0-4bfb-bb9f-889636a4a9fa" Oct 02 09:48:44 crc kubenswrapper[4771]: I1002 09:48:44.681245 4771 scope.go:117] "RemoveContainer" containerID="c35db692e359d56ccbb79957ddfb37fbf56b3ad4e46093fa55f540da775d2441" Oct 02 09:48:45 crc kubenswrapper[4771]: I1002 09:48:45.646641 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xh7xv_bb6b65c1-83b7-4b23-abe8-c4255c6e59f8/kube-multus/2.log" Oct 02 09:48:45 crc kubenswrapper[4771]: I1002 09:48:45.646999 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xh7xv" event={"ID":"bb6b65c1-83b7-4b23-abe8-c4255c6e59f8","Type":"ContainerStarted","Data":"e81a0fb148f2af7d605a7501010b2e7826b6e41ca2c598eb426c3afca4f6c627"} Oct 02 09:48:50 crc kubenswrapper[4771]: I1002 09:48:50.684460 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2drsv" Oct 02 09:48:51 crc kubenswrapper[4771]: I1002 09:48:51.681211 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:51 crc kubenswrapper[4771]: I1002 09:48:51.682218 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" Oct 02 09:48:51 crc kubenswrapper[4771]: I1002 09:48:51.928502 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps"] Oct 02 09:48:52 crc kubenswrapper[4771]: I1002 09:48:52.687919 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" event={"ID":"9c0d8940-6aa3-461a-9b29-53277779bb47","Type":"ContainerStarted","Data":"4f27524507e553ff7ea00864d798247502e7ac17d3b77fac7d349429048c4850"} Oct 02 09:48:53 crc kubenswrapper[4771]: I1002 09:48:53.680678 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:53 crc kubenswrapper[4771]: I1002 09:48:53.680680 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:53 crc kubenswrapper[4771]: I1002 09:48:53.685792 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:53 crc kubenswrapper[4771]: I1002 09:48:53.685862 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" Oct 02 09:48:54 crc kubenswrapper[4771]: I1002 09:48:54.131001 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-2ccmn"] Oct 02 09:48:54 crc kubenswrapper[4771]: I1002 09:48:54.139297 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn"] Oct 02 09:48:54 crc kubenswrapper[4771]: W1002 09:48:54.140407 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa786166_e5d7_43e7_8803_98135ffc2871.slice/crio-fa1331a43ed34163ef887316131a6e626e0baf3dabcd74ac142b35ed0e930410 WatchSource:0}: Error finding container fa1331a43ed34163ef887316131a6e626e0baf3dabcd74ac142b35ed0e930410: Status 404 returned error can't find the container with id fa1331a43ed34163ef887316131a6e626e0baf3dabcd74ac142b35ed0e930410 Oct 02 09:48:54 crc kubenswrapper[4771]: I1002 09:48:54.700201 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" event={"ID":"8fd0bf98-7976-4e81-8afa-76b37aff3944","Type":"ContainerStarted","Data":"1fea7e606931bb91df7796c1cc4b8dca7b7236b90c479af99969f5639041e6e1"} Oct 02 09:48:54 crc kubenswrapper[4771]: I1002 09:48:54.702447 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" event={"ID":"fa786166-e5d7-43e7-8803-98135ffc2871","Type":"ContainerStarted","Data":"fa1331a43ed34163ef887316131a6e626e0baf3dabcd74ac142b35ed0e930410"} Oct 02 09:48:56 crc kubenswrapper[4771]: I1002 09:48:56.680545 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:56 crc kubenswrapper[4771]: I1002 09:48:56.680696 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:56 crc kubenswrapper[4771]: I1002 09:48:56.681491 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:48:56 crc kubenswrapper[4771]: I1002 09:48:56.681654 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" Oct 02 09:48:58 crc kubenswrapper[4771]: I1002 09:48:58.332316 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-wwgrl"] Oct 02 09:48:58 crc kubenswrapper[4771]: I1002 09:48:58.812354 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2"] Oct 02 09:48:59 crc kubenswrapper[4771]: W1002 09:48:59.028967 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6214e26e_87c0_4bfb_bb9f_889636a4a9fa.slice/crio-d61e155d25d8612ee21b90a230fcba6d33f0188a70ed193d0d0061027f2daac3 WatchSource:0}: Error finding container d61e155d25d8612ee21b90a230fcba6d33f0188a70ed193d0d0061027f2daac3: Status 404 returned error can't find the container with id d61e155d25d8612ee21b90a230fcba6d33f0188a70ed193d0d0061027f2daac3 Oct 02 09:48:59 crc kubenswrapper[4771]: I1002 09:48:59.744731 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" event={"ID":"e792e389-a48d-47fa-8c41-9767cb669865","Type":"ContainerStarted","Data":"1ab17c4ea82d0b444f5ab51db252e4e38be4450779c35426e9a228e6ceb2b370"} Oct 02 09:48:59 crc kubenswrapper[4771]: I1002 09:48:59.745849 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" event={"ID":"6214e26e-87c0-4bfb-bb9f-889636a4a9fa","Type":"ContainerStarted","Data":"d61e155d25d8612ee21b90a230fcba6d33f0188a70ed193d0d0061027f2daac3"} Oct 02 09:48:59 crc kubenswrapper[4771]: I1002 09:48:59.747188 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" event={"ID":"9c0d8940-6aa3-461a-9b29-53277779bb47","Type":"ContainerStarted","Data":"bd74ab483581b81b191aa5302d38268e46083732ea625cbb431e0d32b54d244d"} Oct 02 09:48:59 crc kubenswrapper[4771]: I1002 09:48:59.748574 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" event={"ID":"8fd0bf98-7976-4e81-8afa-76b37aff3944","Type":"ContainerStarted","Data":"87811c7195e01e6f8c0a8fb027fa32ac32f2e9f61bea6af8f8d3b0ae04741091"} Oct 02 09:48:59 crc kubenswrapper[4771]: I1002 09:48:59.749870 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" event={"ID":"fa786166-e5d7-43e7-8803-98135ffc2871","Type":"ContainerStarted","Data":"714c316bc113e858d736344d2a7492660d7d795c8abebe0d4d02abfc9642802a"} Oct 02 09:48:59 crc kubenswrapper[4771]: I1002 09:48:59.750010 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:48:59 crc kubenswrapper[4771]: I1002 09:48:59.768236 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-tbgps" podStartSLOduration=28.650183159 podStartE2EDuration="35.768212246s" podCreationTimestamp="2025-10-02 09:48:24 +0000 UTC" firstStartedPulling="2025-10-02 09:48:51.938266066 +0000 UTC m=+719.585951133" lastFinishedPulling="2025-10-02 09:48:59.056295153 +0000 UTC m=+726.703980220" observedRunningTime="2025-10-02 09:48:59.763383591 +0000 UTC m=+727.411068658" watchObservedRunningTime="2025-10-02 09:48:59.768212246 +0000 UTC m=+727.415897313" Oct 02 09:48:59 crc kubenswrapper[4771]: I1002 09:48:59.781349 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn" podStartSLOduration=30.856721188 podStartE2EDuration="35.781328964s" podCreationTimestamp="2025-10-02 09:48:24 +0000 UTC" firstStartedPulling="2025-10-02 09:48:54.150732329 +0000 UTC m=+721.798417396" lastFinishedPulling="2025-10-02 09:48:59.075340105 +0000 UTC m=+726.723025172" observedRunningTime="2025-10-02 09:48:59.779516488 +0000 UTC m=+727.427201555" watchObservedRunningTime="2025-10-02 09:48:59.781328964 +0000 UTC m=+727.429014031" Oct 02 09:48:59 crc kubenswrapper[4771]: I1002 09:48:59.807377 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" podStartSLOduration=29.886600141 podStartE2EDuration="34.807338637s" podCreationTimestamp="2025-10-02 09:48:25 +0000 UTC" firstStartedPulling="2025-10-02 09:48:54.146805477 +0000 UTC m=+721.794490544" lastFinishedPulling="2025-10-02 09:48:59.067543973 +0000 UTC m=+726.715229040" observedRunningTime="2025-10-02 09:48:59.802332057 +0000 UTC m=+727.450017124" watchObservedRunningTime="2025-10-02 09:48:59.807338637 +0000 UTC m=+727.455023704" Oct 02 09:49:00 crc kubenswrapper[4771]: I1002 09:49:00.763101 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" event={"ID":"6214e26e-87c0-4bfb-bb9f-889636a4a9fa","Type":"ContainerStarted","Data":"0672af893ef2695b5257cf8b958d1be0c92a493207b21dcb051db967ab0db63b"} Oct 02 09:49:00 crc kubenswrapper[4771]: I1002 09:49:00.781670 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2" podStartSLOduration=35.527297932 podStartE2EDuration="36.781648799s" podCreationTimestamp="2025-10-02 09:48:24 +0000 UTC" firstStartedPulling="2025-10-02 09:48:59.058681434 +0000 UTC m=+726.706366501" lastFinishedPulling="2025-10-02 09:49:00.313032301 +0000 UTC m=+727.960717368" observedRunningTime="2025-10-02 09:49:00.778733294 +0000 UTC m=+728.426418371" watchObservedRunningTime="2025-10-02 09:49:00.781648799 +0000 UTC m=+728.429333876" Oct 02 09:49:04 crc kubenswrapper[4771]: I1002 09:49:04.786475 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" event={"ID":"e792e389-a48d-47fa-8c41-9767cb669865","Type":"ContainerStarted","Data":"0df20ac1b39b646a302c4d852344aacaef2fbae582ec1086c84ac07a7a096eed"} Oct 02 09:49:04 crc kubenswrapper[4771]: I1002 09:49:04.787034 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:49:04 crc kubenswrapper[4771]: I1002 09:49:04.788975 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" Oct 02 09:49:04 crc kubenswrapper[4771]: I1002 09:49:04.829760 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-wwgrl" podStartSLOduration=36.005608749 podStartE2EDuration="40.829740837s" podCreationTimestamp="2025-10-02 09:48:24 +0000 UTC" firstStartedPulling="2025-10-02 09:48:59.024145462 +0000 UTC m=+726.671830529" lastFinishedPulling="2025-10-02 09:49:03.84827755 +0000 UTC m=+731.495962617" observedRunningTime="2025-10-02 09:49:04.810736906 +0000 UTC m=+732.458421973" watchObservedRunningTime="2025-10-02 09:49:04.829740837 +0000 UTC m=+732.477425904" Oct 02 09:49:05 crc kubenswrapper[4771]: I1002 09:49:05.370715 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-2ccmn" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.808291 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-z9xqq"] Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.809983 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-z9xqq" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.812939 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.813103 4771 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-kh8nn" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.813530 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-n9xk9"] Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.813917 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.814338 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-n9xk9" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.819758 4771 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-x2sk9" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.829185 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-z9xqq"] Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.834688 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvc2s\" (UniqueName: \"kubernetes.io/projected/1eb44788-9a5a-44d1-9b40-909214b5b5f0-kube-api-access-cvc2s\") pod \"cert-manager-5b446d88c5-n9xk9\" (UID: \"1eb44788-9a5a-44d1-9b40-909214b5b5f0\") " pod="cert-manager/cert-manager-5b446d88c5-n9xk9" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.834819 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9289\" (UniqueName: \"kubernetes.io/projected/0d1d36a5-425d-4694-986f-87a45739053a-kube-api-access-b9289\") pod \"cert-manager-cainjector-7f985d654d-z9xqq\" (UID: \"0d1d36a5-425d-4694-986f-87a45739053a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-z9xqq" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.844297 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-n9xk9"] Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.866006 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rmb84"] Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.867315 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-rmb84" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.873954 4771 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-z9fn7" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.924630 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rmb84"] Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.938288 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8vn2\" (UniqueName: \"kubernetes.io/projected/38e28d25-0d2d-44b8-beb7-e5636e2dfdcd-kube-api-access-l8vn2\") pod \"cert-manager-webhook-5655c58dd6-rmb84\" (UID: \"38e28d25-0d2d-44b8-beb7-e5636e2dfdcd\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rmb84" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.938472 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvc2s\" (UniqueName: \"kubernetes.io/projected/1eb44788-9a5a-44d1-9b40-909214b5b5f0-kube-api-access-cvc2s\") pod \"cert-manager-5b446d88c5-n9xk9\" (UID: \"1eb44788-9a5a-44d1-9b40-909214b5b5f0\") " pod="cert-manager/cert-manager-5b446d88c5-n9xk9" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.938585 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9289\" (UniqueName: \"kubernetes.io/projected/0d1d36a5-425d-4694-986f-87a45739053a-kube-api-access-b9289\") pod \"cert-manager-cainjector-7f985d654d-z9xqq\" (UID: \"0d1d36a5-425d-4694-986f-87a45739053a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-z9xqq" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.960257 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvc2s\" (UniqueName: \"kubernetes.io/projected/1eb44788-9a5a-44d1-9b40-909214b5b5f0-kube-api-access-cvc2s\") pod \"cert-manager-5b446d88c5-n9xk9\" (UID: \"1eb44788-9a5a-44d1-9b40-909214b5b5f0\") " pod="cert-manager/cert-manager-5b446d88c5-n9xk9" Oct 02 09:49:11 crc kubenswrapper[4771]: I1002 09:49:11.976608 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9289\" (UniqueName: \"kubernetes.io/projected/0d1d36a5-425d-4694-986f-87a45739053a-kube-api-access-b9289\") pod \"cert-manager-cainjector-7f985d654d-z9xqq\" (UID: \"0d1d36a5-425d-4694-986f-87a45739053a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-z9xqq" Oct 02 09:49:12 crc kubenswrapper[4771]: I1002 09:49:12.040081 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8vn2\" (UniqueName: \"kubernetes.io/projected/38e28d25-0d2d-44b8-beb7-e5636e2dfdcd-kube-api-access-l8vn2\") pod \"cert-manager-webhook-5655c58dd6-rmb84\" (UID: \"38e28d25-0d2d-44b8-beb7-e5636e2dfdcd\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rmb84" Oct 02 09:49:12 crc kubenswrapper[4771]: I1002 09:49:12.060024 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8vn2\" (UniqueName: \"kubernetes.io/projected/38e28d25-0d2d-44b8-beb7-e5636e2dfdcd-kube-api-access-l8vn2\") pod \"cert-manager-webhook-5655c58dd6-rmb84\" (UID: \"38e28d25-0d2d-44b8-beb7-e5636e2dfdcd\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rmb84" Oct 02 09:49:12 crc kubenswrapper[4771]: I1002 09:49:12.133677 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-z9xqq" Oct 02 09:49:12 crc kubenswrapper[4771]: I1002 09:49:12.146296 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:49:12 crc kubenswrapper[4771]: I1002 09:49:12.146350 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:49:12 crc kubenswrapper[4771]: I1002 09:49:12.153117 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-n9xk9" Oct 02 09:49:12 crc kubenswrapper[4771]: I1002 09:49:12.225859 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-rmb84" Oct 02 09:49:12 crc kubenswrapper[4771]: I1002 09:49:12.392817 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-n9xk9"] Oct 02 09:49:12 crc kubenswrapper[4771]: W1002 09:49:12.412252 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eb44788_9a5a_44d1_9b40_909214b5b5f0.slice/crio-bca31ed36dc7539cfe56a4bcf22f18878889dae2d031a1926193cfa5f80743fb WatchSource:0}: Error finding container bca31ed36dc7539cfe56a4bcf22f18878889dae2d031a1926193cfa5f80743fb: Status 404 returned error can't find the container with id bca31ed36dc7539cfe56a4bcf22f18878889dae2d031a1926193cfa5f80743fb Oct 02 09:49:12 crc kubenswrapper[4771]: I1002 09:49:12.423769 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-z9xqq"] Oct 02 09:49:12 crc kubenswrapper[4771]: W1002 09:49:12.435400 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d1d36a5_425d_4694_986f_87a45739053a.slice/crio-91cac8e44decdfc9a2daaf208b5de72d8a56b5ded727985e2b9755379b992ecd WatchSource:0}: Error finding container 91cac8e44decdfc9a2daaf208b5de72d8a56b5ded727985e2b9755379b992ecd: Status 404 returned error can't find the container with id 91cac8e44decdfc9a2daaf208b5de72d8a56b5ded727985e2b9755379b992ecd Oct 02 09:49:12 crc kubenswrapper[4771]: I1002 09:49:12.509606 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rmb84"] Oct 02 09:49:12 crc kubenswrapper[4771]: W1002 09:49:12.517692 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38e28d25_0d2d_44b8_beb7_e5636e2dfdcd.slice/crio-c7a33dd22d8d2ee6ea6dd1cd2a7884d9cb28b661f757d23698e48e6c4b338db1 WatchSource:0}: Error finding container c7a33dd22d8d2ee6ea6dd1cd2a7884d9cb28b661f757d23698e48e6c4b338db1: Status 404 returned error can't find the container with id c7a33dd22d8d2ee6ea6dd1cd2a7884d9cb28b661f757d23698e48e6c4b338db1 Oct 02 09:49:12 crc kubenswrapper[4771]: I1002 09:49:12.845169 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-z9xqq" event={"ID":"0d1d36a5-425d-4694-986f-87a45739053a","Type":"ContainerStarted","Data":"91cac8e44decdfc9a2daaf208b5de72d8a56b5ded727985e2b9755379b992ecd"} Oct 02 09:49:12 crc kubenswrapper[4771]: I1002 09:49:12.847007 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-n9xk9" event={"ID":"1eb44788-9a5a-44d1-9b40-909214b5b5f0","Type":"ContainerStarted","Data":"bca31ed36dc7539cfe56a4bcf22f18878889dae2d031a1926193cfa5f80743fb"} Oct 02 09:49:12 crc kubenswrapper[4771]: I1002 09:49:12.848340 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-rmb84" event={"ID":"38e28d25-0d2d-44b8-beb7-e5636e2dfdcd","Type":"ContainerStarted","Data":"c7a33dd22d8d2ee6ea6dd1cd2a7884d9cb28b661f757d23698e48e6c4b338db1"} Oct 02 09:49:16 crc kubenswrapper[4771]: I1002 09:49:16.883043 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-n9xk9" event={"ID":"1eb44788-9a5a-44d1-9b40-909214b5b5f0","Type":"ContainerStarted","Data":"48dcf93e25aa9c192803eaaeea84c757be0c10ae476a653612f6574c41a7c6b1"} Oct 02 09:49:16 crc kubenswrapper[4771]: I1002 09:49:16.884865 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-rmb84" event={"ID":"38e28d25-0d2d-44b8-beb7-e5636e2dfdcd","Type":"ContainerStarted","Data":"addae500dbef517d70b1e68c94a3962f3e0e0603c403a5c78d9b0b7e480a2c2f"} Oct 02 09:49:16 crc kubenswrapper[4771]: I1002 09:49:16.885006 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-rmb84" Oct 02 09:49:16 crc kubenswrapper[4771]: I1002 09:49:16.886434 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-z9xqq" event={"ID":"0d1d36a5-425d-4694-986f-87a45739053a","Type":"ContainerStarted","Data":"2b7825e6760cf957761fd2b3bec8034e3052d4f72592b504111b98a23fe16c70"} Oct 02 09:49:16 crc kubenswrapper[4771]: I1002 09:49:16.901378 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-n9xk9" podStartSLOduration=2.318630702 podStartE2EDuration="5.901360938s" podCreationTimestamp="2025-10-02 09:49:11 +0000 UTC" firstStartedPulling="2025-10-02 09:49:12.414388019 +0000 UTC m=+740.062073086" lastFinishedPulling="2025-10-02 09:49:15.997118255 +0000 UTC m=+743.644803322" observedRunningTime="2025-10-02 09:49:16.898186976 +0000 UTC m=+744.545872043" watchObservedRunningTime="2025-10-02 09:49:16.901360938 +0000 UTC m=+744.549046005" Oct 02 09:49:16 crc kubenswrapper[4771]: I1002 09:49:16.925465 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-rmb84" podStartSLOduration=2.437549915 podStartE2EDuration="5.92544393s" podCreationTimestamp="2025-10-02 09:49:11 +0000 UTC" firstStartedPulling="2025-10-02 09:49:12.52046609 +0000 UTC m=+740.168151157" lastFinishedPulling="2025-10-02 09:49:16.008360105 +0000 UTC m=+743.656045172" observedRunningTime="2025-10-02 09:49:16.915518323 +0000 UTC m=+744.563203400" watchObservedRunningTime="2025-10-02 09:49:16.92544393 +0000 UTC m=+744.573128997" Oct 02 09:49:16 crc kubenswrapper[4771]: I1002 09:49:16.938875 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-z9xqq" podStartSLOduration=2.379175996 podStartE2EDuration="5.938854216s" podCreationTimestamp="2025-10-02 09:49:11 +0000 UTC" firstStartedPulling="2025-10-02 09:49:12.437841015 +0000 UTC m=+740.085526082" lastFinishedPulling="2025-10-02 09:49:15.997519225 +0000 UTC m=+743.645204302" observedRunningTime="2025-10-02 09:49:16.936350842 +0000 UTC m=+744.584035919" watchObservedRunningTime="2025-10-02 09:49:16.938854216 +0000 UTC m=+744.586539283" Oct 02 09:49:22 crc kubenswrapper[4771]: I1002 09:49:22.229718 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-rmb84" Oct 02 09:49:23 crc kubenswrapper[4771]: I1002 09:49:23.895624 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-p65v7"] Oct 02 09:49:23 crc kubenswrapper[4771]: I1002 09:49:23.902171 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" podUID="b8f91205-c6e3-4f73-ac1b-21a7e79620c1" containerName="controller-manager" containerID="cri-o://ac526b310389f32d0d8b4b4cd84b4e0fb6a37b70c497f1a4d1f96ab23e183146" gracePeriod=30 Oct 02 09:49:23 crc kubenswrapper[4771]: I1002 09:49:23.988468 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn"] Oct 02 09:49:23 crc kubenswrapper[4771]: I1002 09:49:23.988741 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" podUID="dd438f5d-c3ac-4a51-bc47-acd68fcc356a" containerName="route-controller-manager" containerID="cri-o://8effe7c3b02f19f78c7bb8db02eb76189037b19b40b618504ad04aaea700cc6a" gracePeriod=30 Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.437083 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.445540 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.563510 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgmmt\" (UniqueName: \"kubernetes.io/projected/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-kube-api-access-bgmmt\") pod \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.563581 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-client-ca\") pod \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.563612 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-proxy-ca-bundles\") pod \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.563681 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-client-ca\") pod \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.563738 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-serving-cert\") pod \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.563781 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-config\") pod \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\" (UID: \"b8f91205-c6e3-4f73-ac1b-21a7e79620c1\") " Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.563816 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49xvl\" (UniqueName: \"kubernetes.io/projected/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-kube-api-access-49xvl\") pod \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.563861 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-serving-cert\") pod \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.563883 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-config\") pod \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\" (UID: \"dd438f5d-c3ac-4a51-bc47-acd68fcc356a\") " Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.564876 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-client-ca" (OuterVolumeSpecName: "client-ca") pod "b8f91205-c6e3-4f73-ac1b-21a7e79620c1" (UID: "b8f91205-c6e3-4f73-ac1b-21a7e79620c1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.564976 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-config" (OuterVolumeSpecName: "config") pod "b8f91205-c6e3-4f73-ac1b-21a7e79620c1" (UID: "b8f91205-c6e3-4f73-ac1b-21a7e79620c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.565006 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-config" (OuterVolumeSpecName: "config") pod "dd438f5d-c3ac-4a51-bc47-acd68fcc356a" (UID: "dd438f5d-c3ac-4a51-bc47-acd68fcc356a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.565215 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b8f91205-c6e3-4f73-ac1b-21a7e79620c1" (UID: "b8f91205-c6e3-4f73-ac1b-21a7e79620c1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.565864 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-client-ca" (OuterVolumeSpecName: "client-ca") pod "dd438f5d-c3ac-4a51-bc47-acd68fcc356a" (UID: "dd438f5d-c3ac-4a51-bc47-acd68fcc356a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.571031 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-kube-api-access-bgmmt" (OuterVolumeSpecName: "kube-api-access-bgmmt") pod "b8f91205-c6e3-4f73-ac1b-21a7e79620c1" (UID: "b8f91205-c6e3-4f73-ac1b-21a7e79620c1"). InnerVolumeSpecName "kube-api-access-bgmmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.571532 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-kube-api-access-49xvl" (OuterVolumeSpecName: "kube-api-access-49xvl") pod "dd438f5d-c3ac-4a51-bc47-acd68fcc356a" (UID: "dd438f5d-c3ac-4a51-bc47-acd68fcc356a"). InnerVolumeSpecName "kube-api-access-49xvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.571914 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dd438f5d-c3ac-4a51-bc47-acd68fcc356a" (UID: "dd438f5d-c3ac-4a51-bc47-acd68fcc356a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.574084 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b8f91205-c6e3-4f73-ac1b-21a7e79620c1" (UID: "b8f91205-c6e3-4f73-ac1b-21a7e79620c1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.665871 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgmmt\" (UniqueName: \"kubernetes.io/projected/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-kube-api-access-bgmmt\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.665968 4771 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.665984 4771 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.665996 4771 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.666007 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.666018 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8f91205-c6e3-4f73-ac1b-21a7e79620c1-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.666030 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49xvl\" (UniqueName: \"kubernetes.io/projected/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-kube-api-access-49xvl\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.666045 4771 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.666057 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd438f5d-c3ac-4a51-bc47-acd68fcc356a-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.932913 4771 generic.go:334] "Generic (PLEG): container finished" podID="dd438f5d-c3ac-4a51-bc47-acd68fcc356a" containerID="8effe7c3b02f19f78c7bb8db02eb76189037b19b40b618504ad04aaea700cc6a" exitCode=0 Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.932989 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" event={"ID":"dd438f5d-c3ac-4a51-bc47-acd68fcc356a","Type":"ContainerDied","Data":"8effe7c3b02f19f78c7bb8db02eb76189037b19b40b618504ad04aaea700cc6a"} Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.933020 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" event={"ID":"dd438f5d-c3ac-4a51-bc47-acd68fcc356a","Type":"ContainerDied","Data":"4af5fcd51481b37ef71a69287f3a015d4b7226cf7fae0740a768d39dfb66524f"} Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.933040 4771 scope.go:117] "RemoveContainer" containerID="8effe7c3b02f19f78c7bb8db02eb76189037b19b40b618504ad04aaea700cc6a" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.933160 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.942185 4771 generic.go:334] "Generic (PLEG): container finished" podID="b8f91205-c6e3-4f73-ac1b-21a7e79620c1" containerID="ac526b310389f32d0d8b4b4cd84b4e0fb6a37b70c497f1a4d1f96ab23e183146" exitCode=0 Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.942427 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" event={"ID":"b8f91205-c6e3-4f73-ac1b-21a7e79620c1","Type":"ContainerDied","Data":"ac526b310389f32d0d8b4b4cd84b4e0fb6a37b70c497f1a4d1f96ab23e183146"} Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.942539 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" event={"ID":"b8f91205-c6e3-4f73-ac1b-21a7e79620c1","Type":"ContainerDied","Data":"620cea6fc4f81c9aa4602d0967a02256bdc0bdf3f69d0d3a84103a2b36958296"} Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.942656 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-p65v7" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.973204 4771 scope.go:117] "RemoveContainer" containerID="8effe7c3b02f19f78c7bb8db02eb76189037b19b40b618504ad04aaea700cc6a" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.984462 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn"] Oct 02 09:49:24 crc kubenswrapper[4771]: E1002 09:49:24.989276 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8effe7c3b02f19f78c7bb8db02eb76189037b19b40b618504ad04aaea700cc6a\": container with ID starting with 8effe7c3b02f19f78c7bb8db02eb76189037b19b40b618504ad04aaea700cc6a not found: ID does not exist" containerID="8effe7c3b02f19f78c7bb8db02eb76189037b19b40b618504ad04aaea700cc6a" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.989341 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8effe7c3b02f19f78c7bb8db02eb76189037b19b40b618504ad04aaea700cc6a"} err="failed to get container status \"8effe7c3b02f19f78c7bb8db02eb76189037b19b40b618504ad04aaea700cc6a\": rpc error: code = NotFound desc = could not find container \"8effe7c3b02f19f78c7bb8db02eb76189037b19b40b618504ad04aaea700cc6a\": container with ID starting with 8effe7c3b02f19f78c7bb8db02eb76189037b19b40b618504ad04aaea700cc6a not found: ID does not exist" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.989390 4771 scope.go:117] "RemoveContainer" containerID="ac526b310389f32d0d8b4b4cd84b4e0fb6a37b70c497f1a4d1f96ab23e183146" Oct 02 09:49:24 crc kubenswrapper[4771]: I1002 09:49:24.996274 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g92wn"] Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.009103 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-p65v7"] Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.011337 4771 scope.go:117] "RemoveContainer" containerID="ac526b310389f32d0d8b4b4cd84b4e0fb6a37b70c497f1a4d1f96ab23e183146" Oct 02 09:49:25 crc kubenswrapper[4771]: E1002 09:49:25.011885 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac526b310389f32d0d8b4b4cd84b4e0fb6a37b70c497f1a4d1f96ab23e183146\": container with ID starting with ac526b310389f32d0d8b4b4cd84b4e0fb6a37b70c497f1a4d1f96ab23e183146 not found: ID does not exist" containerID="ac526b310389f32d0d8b4b4cd84b4e0fb6a37b70c497f1a4d1f96ab23e183146" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.011930 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac526b310389f32d0d8b4b4cd84b4e0fb6a37b70c497f1a4d1f96ab23e183146"} err="failed to get container status \"ac526b310389f32d0d8b4b4cd84b4e0fb6a37b70c497f1a4d1f96ab23e183146\": rpc error: code = NotFound desc = could not find container \"ac526b310389f32d0d8b4b4cd84b4e0fb6a37b70c497f1a4d1f96ab23e183146\": container with ID starting with ac526b310389f32d0d8b4b4cd84b4e0fb6a37b70c497f1a4d1f96ab23e183146 not found: ID does not exist" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.018815 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-p65v7"] Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.688850 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8f91205-c6e3-4f73-ac1b-21a7e79620c1" path="/var/lib/kubelet/pods/b8f91205-c6e3-4f73-ac1b-21a7e79620c1/volumes" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.689893 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd438f5d-c3ac-4a51-bc47-acd68fcc356a" path="/var/lib/kubelet/pods/dd438f5d-c3ac-4a51-bc47-acd68fcc356a/volumes" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.775456 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd"] Oct 02 09:49:25 crc kubenswrapper[4771]: E1002 09:49:25.775739 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd438f5d-c3ac-4a51-bc47-acd68fcc356a" containerName="route-controller-manager" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.775754 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd438f5d-c3ac-4a51-bc47-acd68fcc356a" containerName="route-controller-manager" Oct 02 09:49:25 crc kubenswrapper[4771]: E1002 09:49:25.775771 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8f91205-c6e3-4f73-ac1b-21a7e79620c1" containerName="controller-manager" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.775779 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8f91205-c6e3-4f73-ac1b-21a7e79620c1" containerName="controller-manager" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.775923 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8f91205-c6e3-4f73-ac1b-21a7e79620c1" containerName="controller-manager" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.775946 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd438f5d-c3ac-4a51-bc47-acd68fcc356a" containerName="route-controller-manager" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.776441 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.779547 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.779922 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.780115 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.780402 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.780573 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.780753 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.781719 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-87f8cccff-xgz7k"] Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.783048 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.787087 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.787654 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.787806 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.787931 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.788252 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.788285 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.793680 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd"] Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.798527 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-87f8cccff-xgz7k"] Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.803856 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.883680 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15cdbb96-f8e5-45c1-be0a-58e9dab78017-serving-cert\") pod \"route-controller-manager-5674896fb6-nb9pd\" (UID: \"15cdbb96-f8e5-45c1-be0a-58e9dab78017\") " pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.883727 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-client-ca\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.883748 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-config\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.883943 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15cdbb96-f8e5-45c1-be0a-58e9dab78017-config\") pod \"route-controller-manager-5674896fb6-nb9pd\" (UID: \"15cdbb96-f8e5-45c1-be0a-58e9dab78017\") " pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.884054 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-proxy-ca-bundles\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.884089 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-serving-cert\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.884224 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrvd4\" (UniqueName: \"kubernetes.io/projected/15cdbb96-f8e5-45c1-be0a-58e9dab78017-kube-api-access-hrvd4\") pod \"route-controller-manager-5674896fb6-nb9pd\" (UID: \"15cdbb96-f8e5-45c1-be0a-58e9dab78017\") " pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.884398 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/15cdbb96-f8e5-45c1-be0a-58e9dab78017-client-ca\") pod \"route-controller-manager-5674896fb6-nb9pd\" (UID: \"15cdbb96-f8e5-45c1-be0a-58e9dab78017\") " pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.884515 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz9kc\" (UniqueName: \"kubernetes.io/projected/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-kube-api-access-dz9kc\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.985403 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrvd4\" (UniqueName: \"kubernetes.io/projected/15cdbb96-f8e5-45c1-be0a-58e9dab78017-kube-api-access-hrvd4\") pod \"route-controller-manager-5674896fb6-nb9pd\" (UID: \"15cdbb96-f8e5-45c1-be0a-58e9dab78017\") " pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.985482 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/15cdbb96-f8e5-45c1-be0a-58e9dab78017-client-ca\") pod \"route-controller-manager-5674896fb6-nb9pd\" (UID: \"15cdbb96-f8e5-45c1-be0a-58e9dab78017\") " pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.985539 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz9kc\" (UniqueName: \"kubernetes.io/projected/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-kube-api-access-dz9kc\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.985571 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15cdbb96-f8e5-45c1-be0a-58e9dab78017-serving-cert\") pod \"route-controller-manager-5674896fb6-nb9pd\" (UID: \"15cdbb96-f8e5-45c1-be0a-58e9dab78017\") " pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.985592 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-client-ca\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.985614 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-config\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.985654 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15cdbb96-f8e5-45c1-be0a-58e9dab78017-config\") pod \"route-controller-manager-5674896fb6-nb9pd\" (UID: \"15cdbb96-f8e5-45c1-be0a-58e9dab78017\") " pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.985684 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-proxy-ca-bundles\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.985712 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-serving-cert\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.986427 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/15cdbb96-f8e5-45c1-be0a-58e9dab78017-client-ca\") pod \"route-controller-manager-5674896fb6-nb9pd\" (UID: \"15cdbb96-f8e5-45c1-be0a-58e9dab78017\") " pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.986719 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-client-ca\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.987068 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-config\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.987182 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-proxy-ca-bundles\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.988265 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15cdbb96-f8e5-45c1-be0a-58e9dab78017-config\") pod \"route-controller-manager-5674896fb6-nb9pd\" (UID: \"15cdbb96-f8e5-45c1-be0a-58e9dab78017\") " pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.990688 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-serving-cert\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:25 crc kubenswrapper[4771]: I1002 09:49:25.990762 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15cdbb96-f8e5-45c1-be0a-58e9dab78017-serving-cert\") pod \"route-controller-manager-5674896fb6-nb9pd\" (UID: \"15cdbb96-f8e5-45c1-be0a-58e9dab78017\") " pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.006480 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrvd4\" (UniqueName: \"kubernetes.io/projected/15cdbb96-f8e5-45c1-be0a-58e9dab78017-kube-api-access-hrvd4\") pod \"route-controller-manager-5674896fb6-nb9pd\" (UID: \"15cdbb96-f8e5-45c1-be0a-58e9dab78017\") " pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.007755 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz9kc\" (UniqueName: \"kubernetes.io/projected/3597e4e9-ccaf-49c2-b3fb-63f2301fec08-kube-api-access-dz9kc\") pod \"controller-manager-87f8cccff-xgz7k\" (UID: \"3597e4e9-ccaf-49c2-b3fb-63f2301fec08\") " pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.100512 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.109038 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.325663 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-87f8cccff-xgz7k"] Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.391378 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd"] Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.958379 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" event={"ID":"15cdbb96-f8e5-45c1-be0a-58e9dab78017","Type":"ContainerStarted","Data":"d40d834bc79e42ca662e87db64f380fec4da3eca80655428d8a9d58876e31f73"} Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.959160 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" event={"ID":"15cdbb96-f8e5-45c1-be0a-58e9dab78017","Type":"ContainerStarted","Data":"d255fe17350016a549bf31ddabd64d8c65edfaf673b5effe415f0bc0a7389521"} Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.959553 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.960301 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" event={"ID":"3597e4e9-ccaf-49c2-b3fb-63f2301fec08","Type":"ContainerStarted","Data":"b8cf07eae8f3d935bed29ece1836bf635323e0116549037d87face0dbb950d0e"} Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.960330 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" event={"ID":"3597e4e9-ccaf-49c2-b3fb-63f2301fec08","Type":"ContainerStarted","Data":"b0f11d518b94c875ad53b0472d7d56cc6354d1821412e7475aa167a467176856"} Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.960561 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.969051 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" Oct 02 09:49:26 crc kubenswrapper[4771]: I1002 09:49:26.981647 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" podStartSLOduration=2.9816282469999997 podStartE2EDuration="2.981628247s" podCreationTimestamp="2025-10-02 09:49:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:49:26.978479086 +0000 UTC m=+754.626164173" watchObservedRunningTime="2025-10-02 09:49:26.981628247 +0000 UTC m=+754.629313314" Oct 02 09:49:27 crc kubenswrapper[4771]: I1002 09:49:27.006916 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-87f8cccff-xgz7k" podStartSLOduration=3.00690136 podStartE2EDuration="3.00690136s" podCreationTimestamp="2025-10-02 09:49:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:49:27.005211737 +0000 UTC m=+754.652896814" watchObservedRunningTime="2025-10-02 09:49:27.00690136 +0000 UTC m=+754.654586427" Oct 02 09:49:27 crc kubenswrapper[4771]: I1002 09:49:27.038390 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5674896fb6-nb9pd" Oct 02 09:49:30 crc kubenswrapper[4771]: I1002 09:49:30.758321 4771 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 09:49:42 crc kubenswrapper[4771]: I1002 09:49:42.146025 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:49:42 crc kubenswrapper[4771]: I1002 09:49:42.146517 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:49:42 crc kubenswrapper[4771]: I1002 09:49:42.146568 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:49:42 crc kubenswrapper[4771]: I1002 09:49:42.147109 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8f7959ad99e22e4304788ce28b1466fbafe93b57e3295c1b849950830aa10434"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:49:42 crc kubenswrapper[4771]: I1002 09:49:42.147216 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://8f7959ad99e22e4304788ce28b1466fbafe93b57e3295c1b849950830aa10434" gracePeriod=600 Oct 02 09:49:43 crc kubenswrapper[4771]: I1002 09:49:43.065913 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="8f7959ad99e22e4304788ce28b1466fbafe93b57e3295c1b849950830aa10434" exitCode=0 Oct 02 09:49:43 crc kubenswrapper[4771]: I1002 09:49:43.065984 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"8f7959ad99e22e4304788ce28b1466fbafe93b57e3295c1b849950830aa10434"} Oct 02 09:49:43 crc kubenswrapper[4771]: I1002 09:49:43.066276 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"e881357965830ad6d553dbf5d22ee7a11da1ee52ace8a1671edcdf44ed33436b"} Oct 02 09:49:43 crc kubenswrapper[4771]: I1002 09:49:43.066299 4771 scope.go:117] "RemoveContainer" containerID="f2d56a0106e4d97ef1eaacdf9670a1e221241c6eb398fb493028ecc6d5fd6e25" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.516164 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x"] Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.519338 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.521764 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.523864 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x"] Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.587110 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d3c0ef21-3031-4032-89d8-001a64387d57-util\") pod \"b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x\" (UID: \"d3c0ef21-3031-4032-89d8-001a64387d57\") " pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.587231 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klsbn\" (UniqueName: \"kubernetes.io/projected/d3c0ef21-3031-4032-89d8-001a64387d57-kube-api-access-klsbn\") pod \"b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x\" (UID: \"d3c0ef21-3031-4032-89d8-001a64387d57\") " pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.587263 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d3c0ef21-3031-4032-89d8-001a64387d57-bundle\") pod \"b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x\" (UID: \"d3c0ef21-3031-4032-89d8-001a64387d57\") " pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.688626 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d3c0ef21-3031-4032-89d8-001a64387d57-util\") pod \"b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x\" (UID: \"d3c0ef21-3031-4032-89d8-001a64387d57\") " pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.688721 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klsbn\" (UniqueName: \"kubernetes.io/projected/d3c0ef21-3031-4032-89d8-001a64387d57-kube-api-access-klsbn\") pod \"b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x\" (UID: \"d3c0ef21-3031-4032-89d8-001a64387d57\") " pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.688763 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d3c0ef21-3031-4032-89d8-001a64387d57-bundle\") pod \"b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x\" (UID: \"d3c0ef21-3031-4032-89d8-001a64387d57\") " pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.689756 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d3c0ef21-3031-4032-89d8-001a64387d57-util\") pod \"b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x\" (UID: \"d3c0ef21-3031-4032-89d8-001a64387d57\") " pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.690021 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d3c0ef21-3031-4032-89d8-001a64387d57-bundle\") pod \"b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x\" (UID: \"d3c0ef21-3031-4032-89d8-001a64387d57\") " pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.715482 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf"] Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.716625 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.728997 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klsbn\" (UniqueName: \"kubernetes.io/projected/d3c0ef21-3031-4032-89d8-001a64387d57-kube-api-access-klsbn\") pod \"b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x\" (UID: \"d3c0ef21-3031-4032-89d8-001a64387d57\") " pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.729621 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf"] Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.790356 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64cf8242-2c71-4831-b744-679193ae1ec6-util\") pod \"0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf\" (UID: \"64cf8242-2c71-4831-b744-679193ae1ec6\") " pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.790419 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64cf8242-2c71-4831-b744-679193ae1ec6-bundle\") pod \"0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf\" (UID: \"64cf8242-2c71-4831-b744-679193ae1ec6\") " pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.790484 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvgqh\" (UniqueName: \"kubernetes.io/projected/64cf8242-2c71-4831-b744-679193ae1ec6-kube-api-access-vvgqh\") pod \"0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf\" (UID: \"64cf8242-2c71-4831-b744-679193ae1ec6\") " pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.837812 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.891617 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvgqh\" (UniqueName: \"kubernetes.io/projected/64cf8242-2c71-4831-b744-679193ae1ec6-kube-api-access-vvgqh\") pod \"0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf\" (UID: \"64cf8242-2c71-4831-b744-679193ae1ec6\") " pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.891695 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64cf8242-2c71-4831-b744-679193ae1ec6-util\") pod \"0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf\" (UID: \"64cf8242-2c71-4831-b744-679193ae1ec6\") " pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.891725 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64cf8242-2c71-4831-b744-679193ae1ec6-bundle\") pod \"0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf\" (UID: \"64cf8242-2c71-4831-b744-679193ae1ec6\") " pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.892301 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64cf8242-2c71-4831-b744-679193ae1ec6-bundle\") pod \"0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf\" (UID: \"64cf8242-2c71-4831-b744-679193ae1ec6\") " pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.892404 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64cf8242-2c71-4831-b744-679193ae1ec6-util\") pod \"0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf\" (UID: \"64cf8242-2c71-4831-b744-679193ae1ec6\") " pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" Oct 02 09:49:46 crc kubenswrapper[4771]: I1002 09:49:46.915959 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvgqh\" (UniqueName: \"kubernetes.io/projected/64cf8242-2c71-4831-b744-679193ae1ec6-kube-api-access-vvgqh\") pod \"0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf\" (UID: \"64cf8242-2c71-4831-b744-679193ae1ec6\") " pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" Oct 02 09:49:47 crc kubenswrapper[4771]: I1002 09:49:47.066436 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" Oct 02 09:49:47 crc kubenswrapper[4771]: I1002 09:49:47.305865 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x"] Oct 02 09:49:47 crc kubenswrapper[4771]: I1002 09:49:47.522776 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf"] Oct 02 09:49:47 crc kubenswrapper[4771]: W1002 09:49:47.534076 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64cf8242_2c71_4831_b744_679193ae1ec6.slice/crio-a441b021b7639531dc6ad7155192e3ee8e880c67c76f9ac14231aef310fd5425 WatchSource:0}: Error finding container a441b021b7639531dc6ad7155192e3ee8e880c67c76f9ac14231aef310fd5425: Status 404 returned error can't find the container with id a441b021b7639531dc6ad7155192e3ee8e880c67c76f9ac14231aef310fd5425 Oct 02 09:49:48 crc kubenswrapper[4771]: I1002 09:49:48.119057 4771 generic.go:334] "Generic (PLEG): container finished" podID="d3c0ef21-3031-4032-89d8-001a64387d57" containerID="29ef58f457c1db6e4c4065452e1bd4f2826ebed3aff02695e2e1bff375d64320" exitCode=0 Oct 02 09:49:48 crc kubenswrapper[4771]: I1002 09:49:48.119153 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" event={"ID":"d3c0ef21-3031-4032-89d8-001a64387d57","Type":"ContainerDied","Data":"29ef58f457c1db6e4c4065452e1bd4f2826ebed3aff02695e2e1bff375d64320"} Oct 02 09:49:48 crc kubenswrapper[4771]: I1002 09:49:48.119710 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" event={"ID":"d3c0ef21-3031-4032-89d8-001a64387d57","Type":"ContainerStarted","Data":"2ac898f0204635cfdfae01de21da00678f5fc55ce5f451c14cbbaf2e76a504f3"} Oct 02 09:49:48 crc kubenswrapper[4771]: I1002 09:49:48.122446 4771 generic.go:334] "Generic (PLEG): container finished" podID="64cf8242-2c71-4831-b744-679193ae1ec6" containerID="f45f474a577213023aec4a6b76dfdcc38ac0ccc95a3d923b8dba264f7e706824" exitCode=0 Oct 02 09:49:48 crc kubenswrapper[4771]: I1002 09:49:48.122497 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" event={"ID":"64cf8242-2c71-4831-b744-679193ae1ec6","Type":"ContainerDied","Data":"f45f474a577213023aec4a6b76dfdcc38ac0ccc95a3d923b8dba264f7e706824"} Oct 02 09:49:48 crc kubenswrapper[4771]: I1002 09:49:48.122527 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" event={"ID":"64cf8242-2c71-4831-b744-679193ae1ec6","Type":"ContainerStarted","Data":"a441b021b7639531dc6ad7155192e3ee8e880c67c76f9ac14231aef310fd5425"} Oct 02 09:49:50 crc kubenswrapper[4771]: I1002 09:49:50.257775 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wngr9"] Oct 02 09:49:50 crc kubenswrapper[4771]: I1002 09:49:50.262272 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:49:50 crc kubenswrapper[4771]: I1002 09:49:50.268633 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wngr9"] Oct 02 09:49:50 crc kubenswrapper[4771]: I1002 09:49:50.359589 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl2rq\" (UniqueName: \"kubernetes.io/projected/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-kube-api-access-kl2rq\") pod \"redhat-operators-wngr9\" (UID: \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\") " pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:49:50 crc kubenswrapper[4771]: I1002 09:49:50.359697 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-catalog-content\") pod \"redhat-operators-wngr9\" (UID: \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\") " pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:49:50 crc kubenswrapper[4771]: I1002 09:49:50.359720 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-utilities\") pod \"redhat-operators-wngr9\" (UID: \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\") " pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:49:50 crc kubenswrapper[4771]: I1002 09:49:50.460939 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-catalog-content\") pod \"redhat-operators-wngr9\" (UID: \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\") " pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:49:50 crc kubenswrapper[4771]: I1002 09:49:50.461324 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-utilities\") pod \"redhat-operators-wngr9\" (UID: \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\") " pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:49:50 crc kubenswrapper[4771]: I1002 09:49:50.461484 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl2rq\" (UniqueName: \"kubernetes.io/projected/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-kube-api-access-kl2rq\") pod \"redhat-operators-wngr9\" (UID: \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\") " pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:49:50 crc kubenswrapper[4771]: I1002 09:49:50.461563 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-catalog-content\") pod \"redhat-operators-wngr9\" (UID: \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\") " pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:49:50 crc kubenswrapper[4771]: I1002 09:49:50.461737 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-utilities\") pod \"redhat-operators-wngr9\" (UID: \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\") " pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:49:50 crc kubenswrapper[4771]: I1002 09:49:50.485079 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl2rq\" (UniqueName: \"kubernetes.io/projected/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-kube-api-access-kl2rq\") pod \"redhat-operators-wngr9\" (UID: \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\") " pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:49:50 crc kubenswrapper[4771]: I1002 09:49:50.584472 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:49:51 crc kubenswrapper[4771]: I1002 09:49:51.108892 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wngr9"] Oct 02 09:49:51 crc kubenswrapper[4771]: I1002 09:49:51.140901 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wngr9" event={"ID":"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce","Type":"ContainerStarted","Data":"dd759260386334e5c4af56026542385fe01e752a0e3142eb6e31e51720e31e92"} Oct 02 09:49:51 crc kubenswrapper[4771]: I1002 09:49:51.143429 4771 generic.go:334] "Generic (PLEG): container finished" podID="d3c0ef21-3031-4032-89d8-001a64387d57" containerID="975d49ba5b0c20748ad9d6c0303f70f8cd034f8c2a90789fb9332da672a2e552" exitCode=0 Oct 02 09:49:51 crc kubenswrapper[4771]: I1002 09:49:51.143490 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" event={"ID":"d3c0ef21-3031-4032-89d8-001a64387d57","Type":"ContainerDied","Data":"975d49ba5b0c20748ad9d6c0303f70f8cd034f8c2a90789fb9332da672a2e552"} Oct 02 09:49:51 crc kubenswrapper[4771]: I1002 09:49:51.153512 4771 generic.go:334] "Generic (PLEG): container finished" podID="64cf8242-2c71-4831-b744-679193ae1ec6" containerID="7c52f09d1f20ed82098b41caa8b794a6993fa699ca608636fa40bbaf897e45b6" exitCode=0 Oct 02 09:49:51 crc kubenswrapper[4771]: I1002 09:49:51.153565 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" event={"ID":"64cf8242-2c71-4831-b744-679193ae1ec6","Type":"ContainerDied","Data":"7c52f09d1f20ed82098b41caa8b794a6993fa699ca608636fa40bbaf897e45b6"} Oct 02 09:49:52 crc kubenswrapper[4771]: I1002 09:49:52.163769 4771 generic.go:334] "Generic (PLEG): container finished" podID="64cf8242-2c71-4831-b744-679193ae1ec6" containerID="d5f53420ba52b04b9fec21201d1ae60288d50408e732b70b4c7076014763d9f3" exitCode=0 Oct 02 09:49:52 crc kubenswrapper[4771]: I1002 09:49:52.163828 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" event={"ID":"64cf8242-2c71-4831-b744-679193ae1ec6","Type":"ContainerDied","Data":"d5f53420ba52b04b9fec21201d1ae60288d50408e732b70b4c7076014763d9f3"} Oct 02 09:49:52 crc kubenswrapper[4771]: I1002 09:49:52.168170 4771 generic.go:334] "Generic (PLEG): container finished" podID="3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" containerID="5c0cdaa53002b7773b3b81f85ac692f32f859fe938caf2d963fb6a53f89b10d8" exitCode=0 Oct 02 09:49:52 crc kubenswrapper[4771]: I1002 09:49:52.168250 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wngr9" event={"ID":"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce","Type":"ContainerDied","Data":"5c0cdaa53002b7773b3b81f85ac692f32f859fe938caf2d963fb6a53f89b10d8"} Oct 02 09:49:52 crc kubenswrapper[4771]: I1002 09:49:52.170847 4771 generic.go:334] "Generic (PLEG): container finished" podID="d3c0ef21-3031-4032-89d8-001a64387d57" containerID="69fd8ae7a1c2a4da5a55c28a4d269af1f877c649658f8178361bc258ded2e068" exitCode=0 Oct 02 09:49:52 crc kubenswrapper[4771]: I1002 09:49:52.170887 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" event={"ID":"d3c0ef21-3031-4032-89d8-001a64387d57","Type":"ContainerDied","Data":"69fd8ae7a1c2a4da5a55c28a4d269af1f877c649658f8178361bc258ded2e068"} Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.180545 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wngr9" event={"ID":"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce","Type":"ContainerStarted","Data":"42f17eb88ab4744cd2dba0f9f261447b4453706dc3c8ae58b1cdce7dbe817053"} Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.769935 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.777746 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.830403 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvgqh\" (UniqueName: \"kubernetes.io/projected/64cf8242-2c71-4831-b744-679193ae1ec6-kube-api-access-vvgqh\") pod \"64cf8242-2c71-4831-b744-679193ae1ec6\" (UID: \"64cf8242-2c71-4831-b744-679193ae1ec6\") " Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.830458 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d3c0ef21-3031-4032-89d8-001a64387d57-bundle\") pod \"d3c0ef21-3031-4032-89d8-001a64387d57\" (UID: \"d3c0ef21-3031-4032-89d8-001a64387d57\") " Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.830557 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64cf8242-2c71-4831-b744-679193ae1ec6-util\") pod \"64cf8242-2c71-4831-b744-679193ae1ec6\" (UID: \"64cf8242-2c71-4831-b744-679193ae1ec6\") " Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.830609 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64cf8242-2c71-4831-b744-679193ae1ec6-bundle\") pod \"64cf8242-2c71-4831-b744-679193ae1ec6\" (UID: \"64cf8242-2c71-4831-b744-679193ae1ec6\") " Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.830654 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d3c0ef21-3031-4032-89d8-001a64387d57-util\") pod \"d3c0ef21-3031-4032-89d8-001a64387d57\" (UID: \"d3c0ef21-3031-4032-89d8-001a64387d57\") " Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.830680 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klsbn\" (UniqueName: \"kubernetes.io/projected/d3c0ef21-3031-4032-89d8-001a64387d57-kube-api-access-klsbn\") pod \"d3c0ef21-3031-4032-89d8-001a64387d57\" (UID: \"d3c0ef21-3031-4032-89d8-001a64387d57\") " Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.833456 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3c0ef21-3031-4032-89d8-001a64387d57-bundle" (OuterVolumeSpecName: "bundle") pod "d3c0ef21-3031-4032-89d8-001a64387d57" (UID: "d3c0ef21-3031-4032-89d8-001a64387d57"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.834459 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64cf8242-2c71-4831-b744-679193ae1ec6-bundle" (OuterVolumeSpecName: "bundle") pod "64cf8242-2c71-4831-b744-679193ae1ec6" (UID: "64cf8242-2c71-4831-b744-679193ae1ec6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.836470 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3c0ef21-3031-4032-89d8-001a64387d57-kube-api-access-klsbn" (OuterVolumeSpecName: "kube-api-access-klsbn") pod "d3c0ef21-3031-4032-89d8-001a64387d57" (UID: "d3c0ef21-3031-4032-89d8-001a64387d57"). InnerVolumeSpecName "kube-api-access-klsbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.839605 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64cf8242-2c71-4831-b744-679193ae1ec6-kube-api-access-vvgqh" (OuterVolumeSpecName: "kube-api-access-vvgqh") pod "64cf8242-2c71-4831-b744-679193ae1ec6" (UID: "64cf8242-2c71-4831-b744-679193ae1ec6"). InnerVolumeSpecName "kube-api-access-vvgqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.932489 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klsbn\" (UniqueName: \"kubernetes.io/projected/d3c0ef21-3031-4032-89d8-001a64387d57-kube-api-access-klsbn\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.932527 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvgqh\" (UniqueName: \"kubernetes.io/projected/64cf8242-2c71-4831-b744-679193ae1ec6-kube-api-access-vvgqh\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.932544 4771 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d3c0ef21-3031-4032-89d8-001a64387d57-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:53 crc kubenswrapper[4771]: I1002 09:49:53.932556 4771 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64cf8242-2c71-4831-b744-679193ae1ec6-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:54 crc kubenswrapper[4771]: I1002 09:49:54.179346 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64cf8242-2c71-4831-b744-679193ae1ec6-util" (OuterVolumeSpecName: "util") pod "64cf8242-2c71-4831-b744-679193ae1ec6" (UID: "64cf8242-2c71-4831-b744-679193ae1ec6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:54 crc kubenswrapper[4771]: I1002 09:49:54.185938 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3c0ef21-3031-4032-89d8-001a64387d57-util" (OuterVolumeSpecName: "util") pod "d3c0ef21-3031-4032-89d8-001a64387d57" (UID: "d3c0ef21-3031-4032-89d8-001a64387d57"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:54 crc kubenswrapper[4771]: I1002 09:49:54.191992 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" event={"ID":"64cf8242-2c71-4831-b744-679193ae1ec6","Type":"ContainerDied","Data":"a441b021b7639531dc6ad7155192e3ee8e880c67c76f9ac14231aef310fd5425"} Oct 02 09:49:54 crc kubenswrapper[4771]: I1002 09:49:54.192036 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a441b021b7639531dc6ad7155192e3ee8e880c67c76f9ac14231aef310fd5425" Oct 02 09:49:54 crc kubenswrapper[4771]: I1002 09:49:54.192105 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf" Oct 02 09:49:54 crc kubenswrapper[4771]: I1002 09:49:54.198612 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" event={"ID":"d3c0ef21-3031-4032-89d8-001a64387d57","Type":"ContainerDied","Data":"2ac898f0204635cfdfae01de21da00678f5fc55ce5f451c14cbbaf2e76a504f3"} Oct 02 09:49:54 crc kubenswrapper[4771]: I1002 09:49:54.198648 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ac898f0204635cfdfae01de21da00678f5fc55ce5f451c14cbbaf2e76a504f3" Oct 02 09:49:54 crc kubenswrapper[4771]: I1002 09:49:54.198684 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x" Oct 02 09:49:54 crc kubenswrapper[4771]: I1002 09:49:54.238031 4771 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64cf8242-2c71-4831-b744-679193ae1ec6-util\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:54 crc kubenswrapper[4771]: I1002 09:49:54.238077 4771 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d3c0ef21-3031-4032-89d8-001a64387d57-util\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:55 crc kubenswrapper[4771]: I1002 09:49:55.206104 4771 generic.go:334] "Generic (PLEG): container finished" podID="3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" containerID="42f17eb88ab4744cd2dba0f9f261447b4453706dc3c8ae58b1cdce7dbe817053" exitCode=0 Oct 02 09:49:55 crc kubenswrapper[4771]: I1002 09:49:55.206162 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wngr9" event={"ID":"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce","Type":"ContainerDied","Data":"42f17eb88ab4744cd2dba0f9f261447b4453706dc3c8ae58b1cdce7dbe817053"} Oct 02 09:49:56 crc kubenswrapper[4771]: I1002 09:49:56.215485 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wngr9" event={"ID":"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce","Type":"ContainerStarted","Data":"2034a207577c3d75e73f2e0d5fe01172219c2c3351072aaf969ad5b74c004bf9"} Oct 02 09:49:56 crc kubenswrapper[4771]: I1002 09:49:56.237604 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wngr9" podStartSLOduration=2.5264363359999997 podStartE2EDuration="6.23758245s" podCreationTimestamp="2025-10-02 09:49:50 +0000 UTC" firstStartedPulling="2025-10-02 09:49:52.170468924 +0000 UTC m=+779.818153991" lastFinishedPulling="2025-10-02 09:49:55.881615038 +0000 UTC m=+783.529300105" observedRunningTime="2025-10-02 09:49:56.232210098 +0000 UTC m=+783.879895175" watchObservedRunningTime="2025-10-02 09:49:56.23758245 +0000 UTC m=+783.885267517" Oct 02 09:50:00 crc kubenswrapper[4771]: I1002 09:50:00.585080 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:50:00 crc kubenswrapper[4771]: I1002 09:50:00.586096 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:50:01 crc kubenswrapper[4771]: I1002 09:50:01.640277 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wngr9" podUID="3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" containerName="registry-server" probeResult="failure" output=< Oct 02 09:50:01 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 09:50:01 crc kubenswrapper[4771]: > Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.359269 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv"] Oct 02 09:50:04 crc kubenswrapper[4771]: E1002 09:50:04.361074 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64cf8242-2c71-4831-b744-679193ae1ec6" containerName="pull" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.361196 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="64cf8242-2c71-4831-b744-679193ae1ec6" containerName="pull" Oct 02 09:50:04 crc kubenswrapper[4771]: E1002 09:50:04.361273 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3c0ef21-3031-4032-89d8-001a64387d57" containerName="util" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.361450 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3c0ef21-3031-4032-89d8-001a64387d57" containerName="util" Oct 02 09:50:04 crc kubenswrapper[4771]: E1002 09:50:04.361545 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3c0ef21-3031-4032-89d8-001a64387d57" containerName="pull" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.361615 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3c0ef21-3031-4032-89d8-001a64387d57" containerName="pull" Oct 02 09:50:04 crc kubenswrapper[4771]: E1002 09:50:04.361724 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64cf8242-2c71-4831-b744-679193ae1ec6" containerName="extract" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.361802 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="64cf8242-2c71-4831-b744-679193ae1ec6" containerName="extract" Oct 02 09:50:04 crc kubenswrapper[4771]: E1002 09:50:04.361887 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3c0ef21-3031-4032-89d8-001a64387d57" containerName="extract" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.361956 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3c0ef21-3031-4032-89d8-001a64387d57" containerName="extract" Oct 02 09:50:04 crc kubenswrapper[4771]: E1002 09:50:04.362037 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64cf8242-2c71-4831-b744-679193ae1ec6" containerName="util" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.362102 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="64cf8242-2c71-4831-b744-679193ae1ec6" containerName="util" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.362350 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3c0ef21-3031-4032-89d8-001a64387d57" containerName="extract" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.362437 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="64cf8242-2c71-4831-b744-679193ae1ec6" containerName="extract" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.363458 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.367053 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.367142 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-tqlp7" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.367476 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.369365 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.369482 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.369977 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.396596 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv"] Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.414448 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hd6k\" (UniqueName: \"kubernetes.io/projected/063c19c9-6772-482a-bcbe-b1c620aac916-kube-api-access-7hd6k\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.414546 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/063c19c9-6772-482a-bcbe-b1c620aac916-apiservice-cert\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.414601 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/063c19c9-6772-482a-bcbe-b1c620aac916-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.414686 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/063c19c9-6772-482a-bcbe-b1c620aac916-webhook-cert\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.414744 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/063c19c9-6772-482a-bcbe-b1c620aac916-manager-config\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.516377 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/063c19c9-6772-482a-bcbe-b1c620aac916-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.516556 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/063c19c9-6772-482a-bcbe-b1c620aac916-webhook-cert\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.516671 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/063c19c9-6772-482a-bcbe-b1c620aac916-manager-config\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.516713 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hd6k\" (UniqueName: \"kubernetes.io/projected/063c19c9-6772-482a-bcbe-b1c620aac916-kube-api-access-7hd6k\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.516747 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/063c19c9-6772-482a-bcbe-b1c620aac916-apiservice-cert\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.518162 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/063c19c9-6772-482a-bcbe-b1c620aac916-manager-config\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.528895 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/063c19c9-6772-482a-bcbe-b1c620aac916-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.529154 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/063c19c9-6772-482a-bcbe-b1c620aac916-webhook-cert\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.533802 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/063c19c9-6772-482a-bcbe-b1c620aac916-apiservice-cert\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.554201 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hd6k\" (UniqueName: \"kubernetes.io/projected/063c19c9-6772-482a-bcbe-b1c620aac916-kube-api-access-7hd6k\") pod \"loki-operator-controller-manager-6584497fb7-fcntv\" (UID: \"063c19c9-6772-482a-bcbe-b1c620aac916\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:04 crc kubenswrapper[4771]: I1002 09:50:04.682831 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:05 crc kubenswrapper[4771]: I1002 09:50:05.303391 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv"] Oct 02 09:50:05 crc kubenswrapper[4771]: I1002 09:50:05.862808 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-8958c8b87-r9qsk"] Oct 02 09:50:05 crc kubenswrapper[4771]: I1002 09:50:05.864658 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-8958c8b87-r9qsk" Oct 02 09:50:05 crc kubenswrapper[4771]: I1002 09:50:05.867378 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-bd7m8" Oct 02 09:50:05 crc kubenswrapper[4771]: I1002 09:50:05.868888 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Oct 02 09:50:05 crc kubenswrapper[4771]: I1002 09:50:05.869411 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Oct 02 09:50:05 crc kubenswrapper[4771]: I1002 09:50:05.882364 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-8958c8b87-r9qsk"] Oct 02 09:50:05 crc kubenswrapper[4771]: I1002 09:50:05.969011 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b88g\" (UniqueName: \"kubernetes.io/projected/7d4d0fad-fc61-4f27-b989-be4651af1895-kube-api-access-6b88g\") pod \"cluster-logging-operator-8958c8b87-r9qsk\" (UID: \"7d4d0fad-fc61-4f27-b989-be4651af1895\") " pod="openshift-logging/cluster-logging-operator-8958c8b87-r9qsk" Oct 02 09:50:06 crc kubenswrapper[4771]: I1002 09:50:06.070890 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b88g\" (UniqueName: \"kubernetes.io/projected/7d4d0fad-fc61-4f27-b989-be4651af1895-kube-api-access-6b88g\") pod \"cluster-logging-operator-8958c8b87-r9qsk\" (UID: \"7d4d0fad-fc61-4f27-b989-be4651af1895\") " pod="openshift-logging/cluster-logging-operator-8958c8b87-r9qsk" Oct 02 09:50:06 crc kubenswrapper[4771]: I1002 09:50:06.106904 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b88g\" (UniqueName: \"kubernetes.io/projected/7d4d0fad-fc61-4f27-b989-be4651af1895-kube-api-access-6b88g\") pod \"cluster-logging-operator-8958c8b87-r9qsk\" (UID: \"7d4d0fad-fc61-4f27-b989-be4651af1895\") " pod="openshift-logging/cluster-logging-operator-8958c8b87-r9qsk" Oct 02 09:50:06 crc kubenswrapper[4771]: I1002 09:50:06.188704 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-8958c8b87-r9qsk" Oct 02 09:50:06 crc kubenswrapper[4771]: I1002 09:50:06.292378 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" event={"ID":"063c19c9-6772-482a-bcbe-b1c620aac916","Type":"ContainerStarted","Data":"ff70353cae00b0a8e9ab92f82ba9f3c28197846f00e468090505048e05abedf2"} Oct 02 09:50:06 crc kubenswrapper[4771]: I1002 09:50:06.750855 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-8958c8b87-r9qsk"] Oct 02 09:50:07 crc kubenswrapper[4771]: I1002 09:50:07.300147 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-8958c8b87-r9qsk" event={"ID":"7d4d0fad-fc61-4f27-b989-be4651af1895","Type":"ContainerStarted","Data":"3541bace9d8cc006363ec1d40859228175af1237279fab04c5853a1eb8a11131"} Oct 02 09:50:10 crc kubenswrapper[4771]: I1002 09:50:10.661215 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:50:10 crc kubenswrapper[4771]: I1002 09:50:10.763062 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:50:13 crc kubenswrapper[4771]: I1002 09:50:13.845352 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wngr9"] Oct 02 09:50:13 crc kubenswrapper[4771]: I1002 09:50:13.845876 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wngr9" podUID="3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" containerName="registry-server" containerID="cri-o://2034a207577c3d75e73f2e0d5fe01172219c2c3351072aaf969ad5b74c004bf9" gracePeriod=2 Oct 02 09:50:14 crc kubenswrapper[4771]: I1002 09:50:14.412960 4771 generic.go:334] "Generic (PLEG): container finished" podID="3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" containerID="2034a207577c3d75e73f2e0d5fe01172219c2c3351072aaf969ad5b74c004bf9" exitCode=0 Oct 02 09:50:14 crc kubenswrapper[4771]: I1002 09:50:14.413009 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wngr9" event={"ID":"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce","Type":"ContainerDied","Data":"2034a207577c3d75e73f2e0d5fe01172219c2c3351072aaf969ad5b74c004bf9"} Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.372079 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.445580 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wngr9" event={"ID":"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce","Type":"ContainerDied","Data":"dd759260386334e5c4af56026542385fe01e752a0e3142eb6e31e51720e31e92"} Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.445622 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wngr9" Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.445631 4771 scope.go:117] "RemoveContainer" containerID="2034a207577c3d75e73f2e0d5fe01172219c2c3351072aaf969ad5b74c004bf9" Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.463995 4771 scope.go:117] "RemoveContainer" containerID="42f17eb88ab4744cd2dba0f9f261447b4453706dc3c8ae58b1cdce7dbe817053" Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.478653 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-utilities\") pod \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\" (UID: \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\") " Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.478729 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl2rq\" (UniqueName: \"kubernetes.io/projected/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-kube-api-access-kl2rq\") pod \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\" (UID: \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\") " Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.478787 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-catalog-content\") pod \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\" (UID: \"3d48bdc7-b538-4d35-8827-6d3a5b69d0ce\") " Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.479654 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-utilities" (OuterVolumeSpecName: "utilities") pod "3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" (UID: "3d48bdc7-b538-4d35-8827-6d3a5b69d0ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.480564 4771 scope.go:117] "RemoveContainer" containerID="5c0cdaa53002b7773b3b81f85ac692f32f859fe938caf2d963fb6a53f89b10d8" Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.483305 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-kube-api-access-kl2rq" (OuterVolumeSpecName: "kube-api-access-kl2rq") pod "3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" (UID: "3d48bdc7-b538-4d35-8827-6d3a5b69d0ce"). InnerVolumeSpecName "kube-api-access-kl2rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.565270 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" (UID: "3d48bdc7-b538-4d35-8827-6d3a5b69d0ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.580757 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.580805 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl2rq\" (UniqueName: \"kubernetes.io/projected/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-kube-api-access-kl2rq\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.580818 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.773160 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wngr9"] Oct 02 09:50:16 crc kubenswrapper[4771]: I1002 09:50:16.779194 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wngr9"] Oct 02 09:50:17 crc kubenswrapper[4771]: I1002 09:50:17.453693 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" event={"ID":"063c19c9-6772-482a-bcbe-b1c620aac916","Type":"ContainerStarted","Data":"b9a294bbeefd20d98a42bb5b9bf007bfccbaf4df7cbf11c7981eb2f6c277c8ad"} Oct 02 09:50:17 crc kubenswrapper[4771]: I1002 09:50:17.457421 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-8958c8b87-r9qsk" event={"ID":"7d4d0fad-fc61-4f27-b989-be4651af1895","Type":"ContainerStarted","Data":"0d16c754ad9bd4ee99229f464fe40f056d6882f45889a2f76a1b16a52334ca4c"} Oct 02 09:50:17 crc kubenswrapper[4771]: I1002 09:50:17.478375 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-8958c8b87-r9qsk" podStartSLOduration=3.05050633 podStartE2EDuration="12.478349949s" podCreationTimestamp="2025-10-02 09:50:05 +0000 UTC" firstStartedPulling="2025-10-02 09:50:06.757488217 +0000 UTC m=+794.405173284" lastFinishedPulling="2025-10-02 09:50:16.185331836 +0000 UTC m=+803.833016903" observedRunningTime="2025-10-02 09:50:17.47597133 +0000 UTC m=+805.123656407" watchObservedRunningTime="2025-10-02 09:50:17.478349949 +0000 UTC m=+805.126035026" Oct 02 09:50:17 crc kubenswrapper[4771]: I1002 09:50:17.689457 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" path="/var/lib/kubelet/pods/3d48bdc7-b538-4d35-8827-6d3a5b69d0ce/volumes" Oct 02 09:50:25 crc kubenswrapper[4771]: I1002 09:50:25.512693 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" event={"ID":"063c19c9-6772-482a-bcbe-b1c620aac916","Type":"ContainerStarted","Data":"7db3166c4a3e501d169d6cdb3b3a836a456843e1795512b9c70e13beefe516df"} Oct 02 09:50:25 crc kubenswrapper[4771]: I1002 09:50:25.513268 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:25 crc kubenswrapper[4771]: I1002 09:50:25.515466 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" Oct 02 09:50:25 crc kubenswrapper[4771]: I1002 09:50:25.537985 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-6584497fb7-fcntv" podStartSLOduration=1.636988761 podStartE2EDuration="21.537960178s" podCreationTimestamp="2025-10-02 09:50:04 +0000 UTC" firstStartedPulling="2025-10-02 09:50:05.315709902 +0000 UTC m=+792.963394969" lastFinishedPulling="2025-10-02 09:50:25.216681319 +0000 UTC m=+812.864366386" observedRunningTime="2025-10-02 09:50:25.532164886 +0000 UTC m=+813.179849953" watchObservedRunningTime="2025-10-02 09:50:25.537960178 +0000 UTC m=+813.185645245" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.656612 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vm66q"] Oct 02 09:50:26 crc kubenswrapper[4771]: E1002 09:50:26.657189 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" containerName="extract-utilities" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.657203 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" containerName="extract-utilities" Oct 02 09:50:26 crc kubenswrapper[4771]: E1002 09:50:26.657210 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" containerName="extract-content" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.657216 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" containerName="extract-content" Oct 02 09:50:26 crc kubenswrapper[4771]: E1002 09:50:26.657226 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" containerName="registry-server" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.657232 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" containerName="registry-server" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.657375 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d48bdc7-b538-4d35-8827-6d3a5b69d0ce" containerName="registry-server" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.658273 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.668841 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vm66q"] Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.846077 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98618258-d18c-4bf1-8456-0985f90d893a-utilities\") pod \"community-operators-vm66q\" (UID: \"98618258-d18c-4bf1-8456-0985f90d893a\") " pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.846615 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blh64\" (UniqueName: \"kubernetes.io/projected/98618258-d18c-4bf1-8456-0985f90d893a-kube-api-access-blh64\") pod \"community-operators-vm66q\" (UID: \"98618258-d18c-4bf1-8456-0985f90d893a\") " pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.846846 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98618258-d18c-4bf1-8456-0985f90d893a-catalog-content\") pod \"community-operators-vm66q\" (UID: \"98618258-d18c-4bf1-8456-0985f90d893a\") " pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.948731 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blh64\" (UniqueName: \"kubernetes.io/projected/98618258-d18c-4bf1-8456-0985f90d893a-kube-api-access-blh64\") pod \"community-operators-vm66q\" (UID: \"98618258-d18c-4bf1-8456-0985f90d893a\") " pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.948804 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98618258-d18c-4bf1-8456-0985f90d893a-catalog-content\") pod \"community-operators-vm66q\" (UID: \"98618258-d18c-4bf1-8456-0985f90d893a\") " pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.948879 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98618258-d18c-4bf1-8456-0985f90d893a-utilities\") pod \"community-operators-vm66q\" (UID: \"98618258-d18c-4bf1-8456-0985f90d893a\") " pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.949445 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98618258-d18c-4bf1-8456-0985f90d893a-catalog-content\") pod \"community-operators-vm66q\" (UID: \"98618258-d18c-4bf1-8456-0985f90d893a\") " pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.949475 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98618258-d18c-4bf1-8456-0985f90d893a-utilities\") pod \"community-operators-vm66q\" (UID: \"98618258-d18c-4bf1-8456-0985f90d893a\") " pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.975421 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blh64\" (UniqueName: \"kubernetes.io/projected/98618258-d18c-4bf1-8456-0985f90d893a-kube-api-access-blh64\") pod \"community-operators-vm66q\" (UID: \"98618258-d18c-4bf1-8456-0985f90d893a\") " pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:26 crc kubenswrapper[4771]: I1002 09:50:26.986226 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:27 crc kubenswrapper[4771]: I1002 09:50:27.357618 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vm66q"] Oct 02 09:50:27 crc kubenswrapper[4771]: W1002 09:50:27.363562 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98618258_d18c_4bf1_8456_0985f90d893a.slice/crio-87ed52152199c6c6ad2c322159cd1906dce087b16c1173fc55651e9cfd6ad932 WatchSource:0}: Error finding container 87ed52152199c6c6ad2c322159cd1906dce087b16c1173fc55651e9cfd6ad932: Status 404 returned error can't find the container with id 87ed52152199c6c6ad2c322159cd1906dce087b16c1173fc55651e9cfd6ad932 Oct 02 09:50:27 crc kubenswrapper[4771]: I1002 09:50:27.526200 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm66q" event={"ID":"98618258-d18c-4bf1-8456-0985f90d893a","Type":"ContainerStarted","Data":"87ed52152199c6c6ad2c322159cd1906dce087b16c1173fc55651e9cfd6ad932"} Oct 02 09:50:28 crc kubenswrapper[4771]: I1002 09:50:28.533928 4771 generic.go:334] "Generic (PLEG): container finished" podID="98618258-d18c-4bf1-8456-0985f90d893a" containerID="cfa00909d241c35eab96270ae3b2021ec0ed489d238c9eaa1efd83d5522ab2bd" exitCode=0 Oct 02 09:50:28 crc kubenswrapper[4771]: I1002 09:50:28.533972 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm66q" event={"ID":"98618258-d18c-4bf1-8456-0985f90d893a","Type":"ContainerDied","Data":"cfa00909d241c35eab96270ae3b2021ec0ed489d238c9eaa1efd83d5522ab2bd"} Oct 02 09:50:29 crc kubenswrapper[4771]: I1002 09:50:29.551222 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm66q" event={"ID":"98618258-d18c-4bf1-8456-0985f90d893a","Type":"ContainerStarted","Data":"58ac9ced0014d7e1200fb5688c78f4c7226a9ec8d28574a4429e6e48504316ec"} Oct 02 09:50:30 crc kubenswrapper[4771]: I1002 09:50:30.558952 4771 generic.go:334] "Generic (PLEG): container finished" podID="98618258-d18c-4bf1-8456-0985f90d893a" containerID="58ac9ced0014d7e1200fb5688c78f4c7226a9ec8d28574a4429e6e48504316ec" exitCode=0 Oct 02 09:50:30 crc kubenswrapper[4771]: I1002 09:50:30.559002 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm66q" event={"ID":"98618258-d18c-4bf1-8456-0985f90d893a","Type":"ContainerDied","Data":"58ac9ced0014d7e1200fb5688c78f4c7226a9ec8d28574a4429e6e48504316ec"} Oct 02 09:50:30 crc kubenswrapper[4771]: I1002 09:50:30.827285 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Oct 02 09:50:30 crc kubenswrapper[4771]: I1002 09:50:30.828604 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Oct 02 09:50:30 crc kubenswrapper[4771]: I1002 09:50:30.831065 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Oct 02 09:50:30 crc kubenswrapper[4771]: I1002 09:50:30.831510 4771 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-m5z7q" Oct 02 09:50:30 crc kubenswrapper[4771]: I1002 09:50:30.832523 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Oct 02 09:50:30 crc kubenswrapper[4771]: I1002 09:50:30.834662 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Oct 02 09:50:31 crc kubenswrapper[4771]: I1002 09:50:31.008860 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg4x4\" (UniqueName: \"kubernetes.io/projected/42dd5547-b1c5-422b-8e00-353c72206334-kube-api-access-zg4x4\") pod \"minio\" (UID: \"42dd5547-b1c5-422b-8e00-353c72206334\") " pod="minio-dev/minio" Oct 02 09:50:31 crc kubenswrapper[4771]: I1002 09:50:31.008981 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-162ea0bd-8bd6-4f86-8571-81549963f783\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162ea0bd-8bd6-4f86-8571-81549963f783\") pod \"minio\" (UID: \"42dd5547-b1c5-422b-8e00-353c72206334\") " pod="minio-dev/minio" Oct 02 09:50:31 crc kubenswrapper[4771]: I1002 09:50:31.110586 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg4x4\" (UniqueName: \"kubernetes.io/projected/42dd5547-b1c5-422b-8e00-353c72206334-kube-api-access-zg4x4\") pod \"minio\" (UID: \"42dd5547-b1c5-422b-8e00-353c72206334\") " pod="minio-dev/minio" Oct 02 09:50:31 crc kubenswrapper[4771]: I1002 09:50:31.110679 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-162ea0bd-8bd6-4f86-8571-81549963f783\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162ea0bd-8bd6-4f86-8571-81549963f783\") pod \"minio\" (UID: \"42dd5547-b1c5-422b-8e00-353c72206334\") " pod="minio-dev/minio" Oct 02 09:50:31 crc kubenswrapper[4771]: I1002 09:50:31.113914 4771 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 09:50:31 crc kubenswrapper[4771]: I1002 09:50:31.113953 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-162ea0bd-8bd6-4f86-8571-81549963f783\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162ea0bd-8bd6-4f86-8571-81549963f783\") pod \"minio\" (UID: \"42dd5547-b1c5-422b-8e00-353c72206334\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/432929412a8c45277dfc07b11865ec3ce3dd16fd4cf2756f4f9d2e71c9b45943/globalmount\"" pod="minio-dev/minio" Oct 02 09:50:31 crc kubenswrapper[4771]: I1002 09:50:31.137492 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-162ea0bd-8bd6-4f86-8571-81549963f783\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162ea0bd-8bd6-4f86-8571-81549963f783\") pod \"minio\" (UID: \"42dd5547-b1c5-422b-8e00-353c72206334\") " pod="minio-dev/minio" Oct 02 09:50:31 crc kubenswrapper[4771]: I1002 09:50:31.144520 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg4x4\" (UniqueName: \"kubernetes.io/projected/42dd5547-b1c5-422b-8e00-353c72206334-kube-api-access-zg4x4\") pod \"minio\" (UID: \"42dd5547-b1c5-422b-8e00-353c72206334\") " pod="minio-dev/minio" Oct 02 09:50:31 crc kubenswrapper[4771]: I1002 09:50:31.148825 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Oct 02 09:50:31 crc kubenswrapper[4771]: I1002 09:50:31.572600 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm66q" event={"ID":"98618258-d18c-4bf1-8456-0985f90d893a","Type":"ContainerStarted","Data":"903af8012c4e396f0e6215e9941b260ae0d151a4e56b01a6862ffc42c6038576"} Oct 02 09:50:31 crc kubenswrapper[4771]: I1002 09:50:31.574581 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Oct 02 09:50:32 crc kubenswrapper[4771]: I1002 09:50:32.592397 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"42dd5547-b1c5-422b-8e00-353c72206334","Type":"ContainerStarted","Data":"1dde0998734431c4ee6b27a7723c397299e8910a8304b9f333a3fb3414559bdd"} Oct 02 09:50:33 crc kubenswrapper[4771]: I1002 09:50:33.702550 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vm66q" podStartSLOduration=4.943581974 podStartE2EDuration="7.702531196s" podCreationTimestamp="2025-10-02 09:50:26 +0000 UTC" firstStartedPulling="2025-10-02 09:50:28.54940772 +0000 UTC m=+816.197092787" lastFinishedPulling="2025-10-02 09:50:31.308356942 +0000 UTC m=+818.956042009" observedRunningTime="2025-10-02 09:50:31.600396743 +0000 UTC m=+819.248081810" watchObservedRunningTime="2025-10-02 09:50:33.702531196 +0000 UTC m=+821.350216263" Oct 02 09:50:35 crc kubenswrapper[4771]: I1002 09:50:35.611052 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"42dd5547-b1c5-422b-8e00-353c72206334","Type":"ContainerStarted","Data":"79208347b03cf21dcbb724b316417f3523dc2c7c62e080804d15120067a85f76"} Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.401348 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.719530987 podStartE2EDuration="8.40132869s" podCreationTimestamp="2025-10-02 09:50:28 +0000 UTC" firstStartedPulling="2025-10-02 09:50:31.585183 +0000 UTC m=+819.232868067" lastFinishedPulling="2025-10-02 09:50:35.266980703 +0000 UTC m=+822.914665770" observedRunningTime="2025-10-02 09:50:35.629332492 +0000 UTC m=+823.277017559" watchObservedRunningTime="2025-10-02 09:50:36.40132869 +0000 UTC m=+824.049013757" Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.402648 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6ssks"] Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.403899 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.429435 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6ssks"] Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.519369 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173aa1ff-ba83-466e-b59c-fec9264277e4-catalog-content\") pod \"certified-operators-6ssks\" (UID: \"173aa1ff-ba83-466e-b59c-fec9264277e4\") " pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.519637 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjbpq\" (UniqueName: \"kubernetes.io/projected/173aa1ff-ba83-466e-b59c-fec9264277e4-kube-api-access-sjbpq\") pod \"certified-operators-6ssks\" (UID: \"173aa1ff-ba83-466e-b59c-fec9264277e4\") " pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.519739 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173aa1ff-ba83-466e-b59c-fec9264277e4-utilities\") pod \"certified-operators-6ssks\" (UID: \"173aa1ff-ba83-466e-b59c-fec9264277e4\") " pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.620730 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjbpq\" (UniqueName: \"kubernetes.io/projected/173aa1ff-ba83-466e-b59c-fec9264277e4-kube-api-access-sjbpq\") pod \"certified-operators-6ssks\" (UID: \"173aa1ff-ba83-466e-b59c-fec9264277e4\") " pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.620786 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173aa1ff-ba83-466e-b59c-fec9264277e4-utilities\") pod \"certified-operators-6ssks\" (UID: \"173aa1ff-ba83-466e-b59c-fec9264277e4\") " pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.620815 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173aa1ff-ba83-466e-b59c-fec9264277e4-catalog-content\") pod \"certified-operators-6ssks\" (UID: \"173aa1ff-ba83-466e-b59c-fec9264277e4\") " pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.621410 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173aa1ff-ba83-466e-b59c-fec9264277e4-catalog-content\") pod \"certified-operators-6ssks\" (UID: \"173aa1ff-ba83-466e-b59c-fec9264277e4\") " pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.621693 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173aa1ff-ba83-466e-b59c-fec9264277e4-utilities\") pod \"certified-operators-6ssks\" (UID: \"173aa1ff-ba83-466e-b59c-fec9264277e4\") " pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.647267 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjbpq\" (UniqueName: \"kubernetes.io/projected/173aa1ff-ba83-466e-b59c-fec9264277e4-kube-api-access-sjbpq\") pod \"certified-operators-6ssks\" (UID: \"173aa1ff-ba83-466e-b59c-fec9264277e4\") " pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.721773 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.989043 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:36 crc kubenswrapper[4771]: I1002 09:50:36.989416 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:37 crc kubenswrapper[4771]: I1002 09:50:37.051295 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:37 crc kubenswrapper[4771]: I1002 09:50:37.219668 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6ssks"] Oct 02 09:50:37 crc kubenswrapper[4771]: I1002 09:50:37.627634 4771 generic.go:334] "Generic (PLEG): container finished" podID="173aa1ff-ba83-466e-b59c-fec9264277e4" containerID="a7de4d8e13dc4ce9746c347f4a46369846bc8794a45adf616c69686efadd0d60" exitCode=0 Oct 02 09:50:37 crc kubenswrapper[4771]: I1002 09:50:37.627706 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ssks" event={"ID":"173aa1ff-ba83-466e-b59c-fec9264277e4","Type":"ContainerDied","Data":"a7de4d8e13dc4ce9746c347f4a46369846bc8794a45adf616c69686efadd0d60"} Oct 02 09:50:37 crc kubenswrapper[4771]: I1002 09:50:37.628009 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ssks" event={"ID":"173aa1ff-ba83-466e-b59c-fec9264277e4","Type":"ContainerStarted","Data":"55ca1a4e15e18a496e6f1001bd93f60ff44b0959ca2ea48f11ff7b93428ffd2c"} Oct 02 09:50:37 crc kubenswrapper[4771]: I1002 09:50:37.675098 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.370666 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vm66q"] Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.428808 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj"] Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.429838 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.431593 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.432599 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.432846 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.433709 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-2n9cw" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.435682 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.451579 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj"] Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.568183 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9014359-f9dc-4e80-83c1-127149a3a60e-logging-loki-ca-bundle\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.568240 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h9r2\" (UniqueName: \"kubernetes.io/projected/e9014359-f9dc-4e80-83c1-127149a3a60e-kube-api-access-9h9r2\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.568282 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/e9014359-f9dc-4e80-83c1-127149a3a60e-logging-loki-distributor-http\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.568328 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9014359-f9dc-4e80-83c1-127149a3a60e-config\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.568402 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/e9014359-f9dc-4e80-83c1-127149a3a60e-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.629002 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5d954896cf-74tjf"] Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.629827 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.632463 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.632726 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.634579 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.648080 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5d954896cf-74tjf"] Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.652174 4771 generic.go:334] "Generic (PLEG): container finished" podID="173aa1ff-ba83-466e-b59c-fec9264277e4" containerID="ebffc7b604c96964683ccd5f3fea873620c02a044d237f51ede834a95f9b1a6b" exitCode=0 Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.652279 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ssks" event={"ID":"173aa1ff-ba83-466e-b59c-fec9264277e4","Type":"ContainerDied","Data":"ebffc7b604c96964683ccd5f3fea873620c02a044d237f51ede834a95f9b1a6b"} Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.652395 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vm66q" podUID="98618258-d18c-4bf1-8456-0985f90d893a" containerName="registry-server" containerID="cri-o://903af8012c4e396f0e6215e9941b260ae0d151a4e56b01a6862ffc42c6038576" gracePeriod=2 Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.670056 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9014359-f9dc-4e80-83c1-127149a3a60e-config\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.670120 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/e9014359-f9dc-4e80-83c1-127149a3a60e-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.670244 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9014359-f9dc-4e80-83c1-127149a3a60e-logging-loki-ca-bundle\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.670267 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h9r2\" (UniqueName: \"kubernetes.io/projected/e9014359-f9dc-4e80-83c1-127149a3a60e-kube-api-access-9h9r2\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.670305 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/e9014359-f9dc-4e80-83c1-127149a3a60e-logging-loki-distributor-http\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.672566 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9014359-f9dc-4e80-83c1-127149a3a60e-logging-loki-ca-bundle\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.674013 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9014359-f9dc-4e80-83c1-127149a3a60e-config\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.676851 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/e9014359-f9dc-4e80-83c1-127149a3a60e-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.678573 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/e9014359-f9dc-4e80-83c1-127149a3a60e-logging-loki-distributor-http\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.728233 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h9r2\" (UniqueName: \"kubernetes.io/projected/e9014359-f9dc-4e80-83c1-127149a3a60e-kube-api-access-9h9r2\") pod \"logging-loki-distributor-6f5f7fff97-47hsj\" (UID: \"e9014359-f9dc-4e80-83c1-127149a3a60e\") " pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.736196 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl"] Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.737281 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.744320 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.745236 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.750189 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.765267 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl"] Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.772156 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cr4c\" (UniqueName: \"kubernetes.io/projected/aca1f6a0-c677-4129-84ec-0d33b51193d2-kube-api-access-9cr4c\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.772331 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aca1f6a0-c677-4129-84ec-0d33b51193d2-config\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.772391 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aca1f6a0-c677-4129-84ec-0d33b51193d2-logging-loki-ca-bundle\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.772477 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/aca1f6a0-c677-4129-84ec-0d33b51193d2-logging-loki-querier-grpc\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.772508 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/aca1f6a0-c677-4129-84ec-0d33b51193d2-logging-loki-s3\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.772552 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/aca1f6a0-c677-4129-84ec-0d33b51193d2-logging-loki-querier-http\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.883075 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-config\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.883448 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cr4c\" (UniqueName: \"kubernetes.io/projected/aca1f6a0-c677-4129-84ec-0d33b51193d2-kube-api-access-9cr4c\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.883531 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.883567 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aca1f6a0-c677-4129-84ec-0d33b51193d2-config\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.883601 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.883627 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aca1f6a0-c677-4129-84ec-0d33b51193d2-logging-loki-ca-bundle\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.883653 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.883697 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcxv2\" (UniqueName: \"kubernetes.io/projected/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-kube-api-access-hcxv2\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.883726 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/aca1f6a0-c677-4129-84ec-0d33b51193d2-logging-loki-querier-grpc\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.883758 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/aca1f6a0-c677-4129-84ec-0d33b51193d2-logging-loki-s3\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.883792 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/aca1f6a0-c677-4129-84ec-0d33b51193d2-logging-loki-querier-http\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.885194 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aca1f6a0-c677-4129-84ec-0d33b51193d2-config\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.885800 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aca1f6a0-c677-4129-84ec-0d33b51193d2-logging-loki-ca-bundle\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.891772 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/aca1f6a0-c677-4129-84ec-0d33b51193d2-logging-loki-querier-http\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.893260 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/aca1f6a0-c677-4129-84ec-0d33b51193d2-logging-loki-s3\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.901593 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff"] Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.902229 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/aca1f6a0-c677-4129-84ec-0d33b51193d2-logging-loki-querier-grpc\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.902984 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.906236 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.906292 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.906493 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.906564 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.906667 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.914956 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff"] Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.922615 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cr4c\" (UniqueName: \"kubernetes.io/projected/aca1f6a0-c677-4129-84ec-0d33b51193d2-kube-api-access-9cr4c\") pod \"logging-loki-querier-5d954896cf-74tjf\" (UID: \"aca1f6a0-c677-4129-84ec-0d33b51193d2\") " pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.926657 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh"] Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.930830 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.937520 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-89jmz" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.946565 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.953833 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh"] Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.984591 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.984637 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/21daf834-8ac7-4137-933e-b7d3db27e225-tls-secret\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.984676 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.984703 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21daf834-8ac7-4137-933e-b7d3db27e225-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.984730 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.984760 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21daf834-8ac7-4137-933e-b7d3db27e225-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.985863 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/f9c8afd6-efbc-4787-89ed-100438b03ecd-rbac\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.985899 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcxv2\" (UniqueName: \"kubernetes.io/projected/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-kube-api-access-hcxv2\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.985926 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f9c8afd6-efbc-4787-89ed-100438b03ecd-tls-secret\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.986000 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/f9c8afd6-efbc-4787-89ed-100438b03ecd-tenants\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.986028 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm6np\" (UniqueName: \"kubernetes.io/projected/f9c8afd6-efbc-4787-89ed-100438b03ecd-kube-api-access-cm6np\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.986064 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtcl7\" (UniqueName: \"kubernetes.io/projected/21daf834-8ac7-4137-933e-b7d3db27e225-kube-api-access-vtcl7\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.986105 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/21daf834-8ac7-4137-933e-b7d3db27e225-tenants\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.986155 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/21daf834-8ac7-4137-933e-b7d3db27e225-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.986187 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-config\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.986236 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/f9c8afd6-efbc-4787-89ed-100438b03ecd-lokistack-gateway\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.986262 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/f9c8afd6-efbc-4787-89ed-100438b03ecd-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.986290 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9c8afd6-efbc-4787-89ed-100438b03ecd-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.986321 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/21daf834-8ac7-4137-933e-b7d3db27e225-rbac\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.986355 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/21daf834-8ac7-4137-933e-b7d3db27e225-lokistack-gateway\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.986393 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9c8afd6-efbc-4787-89ed-100438b03ecd-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.987348 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:39 crc kubenswrapper[4771]: I1002 09:50:39.989592 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-config\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:39.994921 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:39.997462 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.017705 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcxv2\" (UniqueName: \"kubernetes.io/projected/cdb0a9ad-c8af-4298-9726-fa21864ee2a6-kube-api-access-hcxv2\") pod \"logging-loki-query-frontend-6fbbbc8b7d-twxfl\" (UID: \"cdb0a9ad-c8af-4298-9726-fa21864ee2a6\") " pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090034 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9c8afd6-efbc-4787-89ed-100438b03ecd-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090102 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/21daf834-8ac7-4137-933e-b7d3db27e225-tls-secret\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090147 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21daf834-8ac7-4137-933e-b7d3db27e225-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090174 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21daf834-8ac7-4137-933e-b7d3db27e225-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090198 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/f9c8afd6-efbc-4787-89ed-100438b03ecd-rbac\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090219 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f9c8afd6-efbc-4787-89ed-100438b03ecd-tls-secret\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090252 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm6np\" (UniqueName: \"kubernetes.io/projected/f9c8afd6-efbc-4787-89ed-100438b03ecd-kube-api-access-cm6np\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090270 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/f9c8afd6-efbc-4787-89ed-100438b03ecd-tenants\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090288 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtcl7\" (UniqueName: \"kubernetes.io/projected/21daf834-8ac7-4137-933e-b7d3db27e225-kube-api-access-vtcl7\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090310 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/21daf834-8ac7-4137-933e-b7d3db27e225-tenants\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090327 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/21daf834-8ac7-4137-933e-b7d3db27e225-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090355 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/f9c8afd6-efbc-4787-89ed-100438b03ecd-lokistack-gateway\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090371 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/f9c8afd6-efbc-4787-89ed-100438b03ecd-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090388 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9c8afd6-efbc-4787-89ed-100438b03ecd-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090435 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/21daf834-8ac7-4137-933e-b7d3db27e225-rbac\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.090455 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/21daf834-8ac7-4137-933e-b7d3db27e225-lokistack-gateway\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.091549 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/21daf834-8ac7-4137-933e-b7d3db27e225-lokistack-gateway\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: E1002 09:50:40.091888 4771 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Oct 02 09:50:40 crc kubenswrapper[4771]: E1002 09:50:40.091965 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/21daf834-8ac7-4137-933e-b7d3db27e225-tls-secret podName:21daf834-8ac7-4137-933e-b7d3db27e225 nodeName:}" failed. No retries permitted until 2025-10-02 09:50:40.59194255 +0000 UTC m=+828.239627707 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/21daf834-8ac7-4137-933e-b7d3db27e225-tls-secret") pod "logging-loki-gateway-6dc97c5fb-2nswh" (UID: "21daf834-8ac7-4137-933e-b7d3db27e225") : secret "logging-loki-gateway-http" not found Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.092499 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/f9c8afd6-efbc-4787-89ed-100438b03ecd-lokistack-gateway\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.093045 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21daf834-8ac7-4137-933e-b7d3db27e225-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.093253 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9c8afd6-efbc-4787-89ed-100438b03ecd-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: E1002 09:50:40.093318 4771 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Oct 02 09:50:40 crc kubenswrapper[4771]: E1002 09:50:40.093359 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f9c8afd6-efbc-4787-89ed-100438b03ecd-tls-secret podName:f9c8afd6-efbc-4787-89ed-100438b03ecd nodeName:}" failed. No retries permitted until 2025-10-02 09:50:40.593344065 +0000 UTC m=+828.241029132 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/f9c8afd6-efbc-4787-89ed-100438b03ecd-tls-secret") pod "logging-loki-gateway-6dc97c5fb-vvdff" (UID: "f9c8afd6-efbc-4787-89ed-100438b03ecd") : secret "logging-loki-gateway-http" not found Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.093589 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/21daf834-8ac7-4137-933e-b7d3db27e225-rbac\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.093728 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21daf834-8ac7-4137-933e-b7d3db27e225-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.093998 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/f9c8afd6-efbc-4787-89ed-100438b03ecd-rbac\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.097736 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/21daf834-8ac7-4137-933e-b7d3db27e225-tenants\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.098669 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/21daf834-8ac7-4137-933e-b7d3db27e225-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.101007 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9c8afd6-efbc-4787-89ed-100438b03ecd-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.103992 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/f9c8afd6-efbc-4787-89ed-100438b03ecd-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.104034 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/f9c8afd6-efbc-4787-89ed-100438b03ecd-tenants\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.113501 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtcl7\" (UniqueName: \"kubernetes.io/projected/21daf834-8ac7-4137-933e-b7d3db27e225-kube-api-access-vtcl7\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.113553 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm6np\" (UniqueName: \"kubernetes.io/projected/f9c8afd6-efbc-4787-89ed-100438b03ecd-kube-api-access-cm6np\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.125569 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.385707 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj"] Oct 02 09:50:40 crc kubenswrapper[4771]: W1002 09:50:40.396911 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9014359_f9dc_4e80_83c1_127149a3a60e.slice/crio-302ce8641e5e4c86b96e696cbf3077b3301cc3ffce7dd7ef90c3778d975b380e WatchSource:0}: Error finding container 302ce8641e5e4c86b96e696cbf3077b3301cc3ffce7dd7ef90c3778d975b380e: Status 404 returned error can't find the container with id 302ce8641e5e4c86b96e696cbf3077b3301cc3ffce7dd7ef90c3778d975b380e Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.415768 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5d954896cf-74tjf"] Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.588176 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.589237 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.591945 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.592145 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.603508 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.607330 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/21daf834-8ac7-4137-933e-b7d3db27e225-tls-secret\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.607410 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f9c8afd6-efbc-4787-89ed-100438b03ecd-tls-secret\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.613222 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f9c8afd6-efbc-4787-89ed-100438b03ecd-tls-secret\") pod \"logging-loki-gateway-6dc97c5fb-vvdff\" (UID: \"f9c8afd6-efbc-4787-89ed-100438b03ecd\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.613559 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/21daf834-8ac7-4137-933e-b7d3db27e225-tls-secret\") pod \"logging-loki-gateway-6dc97c5fb-2nswh\" (UID: \"21daf834-8ac7-4137-933e-b7d3db27e225\") " pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.659895 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" event={"ID":"e9014359-f9dc-4e80-83c1-127149a3a60e","Type":"ContainerStarted","Data":"302ce8641e5e4c86b96e696cbf3077b3301cc3ffce7dd7ef90c3778d975b380e"} Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.661539 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" event={"ID":"aca1f6a0-c677-4129-84ec-0d33b51193d2","Type":"ContainerStarted","Data":"ba5fc57f2b12147a138283a30c9f7e01ab9c4f0de20a761ff40ae8396ce396a6"} Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.702880 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.705530 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.709213 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/34f9375e-2699-41ae-a604-81599cc16ea3-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.709290 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34f9375e-2699-41ae-a604-81599cc16ea3-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.709337 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/34f9375e-2699-41ae-a604-81599cc16ea3-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.709396 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-93738e4f-febd-437e-95b6-c0b3c92a0418\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93738e4f-febd-437e-95b6-c0b3c92a0418\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.709427 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/34f9375e-2699-41ae-a604-81599cc16ea3-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.709452 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34f9375e-2699-41ae-a604-81599cc16ea3-config\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.709474 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1a6b793d-34bf-4984-ab8c-2e67416e545b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a6b793d-34bf-4984-ab8c-2e67416e545b\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.709503 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsd6n\" (UniqueName: \"kubernetes.io/projected/34f9375e-2699-41ae-a604-81599cc16ea3-kube-api-access-lsd6n\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.709646 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.709845 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.721468 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.750495 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl"] Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811186 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811252 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811281 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/34f9375e-2699-41ae-a604-81599cc16ea3-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811314 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34f9375e-2699-41ae-a604-81599cc16ea3-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811385 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqj2d\" (UniqueName: \"kubernetes.io/projected/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-kube-api-access-mqj2d\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811509 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/34f9375e-2699-41ae-a604-81599cc16ea3-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811553 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-93738e4f-febd-437e-95b6-c0b3c92a0418\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93738e4f-febd-437e-95b6-c0b3c92a0418\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811571 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-config\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811595 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/34f9375e-2699-41ae-a604-81599cc16ea3-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811620 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34f9375e-2699-41ae-a604-81599cc16ea3-config\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811637 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-284b5ec6-00ef-43bc-8818-098904e7931e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-284b5ec6-00ef-43bc-8818-098904e7931e\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811659 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1a6b793d-34bf-4984-ab8c-2e67416e545b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a6b793d-34bf-4984-ab8c-2e67416e545b\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811680 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811701 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsd6n\" (UniqueName: \"kubernetes.io/projected/34f9375e-2699-41ae-a604-81599cc16ea3-kube-api-access-lsd6n\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.811719 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.813009 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34f9375e-2699-41ae-a604-81599cc16ea3-config\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.813450 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34f9375e-2699-41ae-a604-81599cc16ea3-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.815937 4771 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.816000 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1a6b793d-34bf-4984-ab8c-2e67416e545b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a6b793d-34bf-4984-ab8c-2e67416e545b\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/36c0a4ca6536079e389ef3b0931cef73f217f8f898cdc8ed58c4390f796dd481/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.816108 4771 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.816228 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-93738e4f-febd-437e-95b6-c0b3c92a0418\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93738e4f-febd-437e-95b6-c0b3c92a0418\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4e71d52f86c1ea0ca503069a4c2e8216a4c948e58a7dadd1d6e38c78c0a0ff64/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.816158 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/34f9375e-2699-41ae-a604-81599cc16ea3-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.816592 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/34f9375e-2699-41ae-a604-81599cc16ea3-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.820974 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/34f9375e-2699-41ae-a604-81599cc16ea3-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.830076 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsd6n\" (UniqueName: \"kubernetes.io/projected/34f9375e-2699-41ae-a604-81599cc16ea3-kube-api-access-lsd6n\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.837473 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.841704 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.842756 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.848013 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-93738e4f-febd-437e-95b6-c0b3c92a0418\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93738e4f-febd-437e-95b6-c0b3c92a0418\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.848864 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1a6b793d-34bf-4984-ab8c-2e67416e545b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a6b793d-34bf-4984-ab8c-2e67416e545b\") pod \"logging-loki-ingester-0\" (UID: \"34f9375e-2699-41ae-a604-81599cc16ea3\") " pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.855086 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.855393 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.873110 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.889500 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.914631 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.915668 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.915745 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.915794 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-00663957-2097-41ec-9a6e-3b4b66bdcb10\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00663957-2097-41ec-9a6e-3b4b66bdcb10\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.915818 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frsz8\" (UniqueName: \"kubernetes.io/projected/c00667ae-ca81-4c00-9ab2-71a1f307978d-kube-api-access-frsz8\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.915855 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/c00667ae-ca81-4c00-9ab2-71a1f307978d-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.915910 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/c00667ae-ca81-4c00-9ab2-71a1f307978d-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.915949 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c00667ae-ca81-4c00-9ab2-71a1f307978d-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.915976 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-config\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.916226 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqj2d\" (UniqueName: \"kubernetes.io/projected/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-kube-api-access-mqj2d\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.916318 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-284b5ec6-00ef-43bc-8818-098904e7931e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-284b5ec6-00ef-43bc-8818-098904e7931e\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.916349 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c00667ae-ca81-4c00-9ab2-71a1f307978d-config\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.916371 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/c00667ae-ca81-4c00-9ab2-71a1f307978d-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.916394 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.916421 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.920160 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.920791 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.921423 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-config\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.921634 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.922322 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.937866 4771 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.938094 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-284b5ec6-00ef-43bc-8818-098904e7931e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-284b5ec6-00ef-43bc-8818-098904e7931e\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0b53b2cf570eed708a6228b693d5af213e3c069e1eece237f0ba966af10e3aaa/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:40 crc kubenswrapper[4771]: I1002 09:50:40.958373 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqj2d\" (UniqueName: \"kubernetes.io/projected/41d40d08-10d1-4abc-90e4-f894d3ccfa4b-kube-api-access-mqj2d\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.017953 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c00667ae-ca81-4c00-9ab2-71a1f307978d-config\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.018001 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/c00667ae-ca81-4c00-9ab2-71a1f307978d-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.018037 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-00663957-2097-41ec-9a6e-3b4b66bdcb10\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00663957-2097-41ec-9a6e-3b4b66bdcb10\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.018058 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frsz8\" (UniqueName: \"kubernetes.io/projected/c00667ae-ca81-4c00-9ab2-71a1f307978d-kube-api-access-frsz8\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.018079 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/c00667ae-ca81-4c00-9ab2-71a1f307978d-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.018107 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/c00667ae-ca81-4c00-9ab2-71a1f307978d-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.018142 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c00667ae-ca81-4c00-9ab2-71a1f307978d-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.019161 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c00667ae-ca81-4c00-9ab2-71a1f307978d-config\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.022736 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c00667ae-ca81-4c00-9ab2-71a1f307978d-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.027318 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/c00667ae-ca81-4c00-9ab2-71a1f307978d-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.030039 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/c00667ae-ca81-4c00-9ab2-71a1f307978d-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.030705 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/c00667ae-ca81-4c00-9ab2-71a1f307978d-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.039319 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frsz8\" (UniqueName: \"kubernetes.io/projected/c00667ae-ca81-4c00-9ab2-71a1f307978d-kube-api-access-frsz8\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.051771 4771 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.051821 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-00663957-2097-41ec-9a6e-3b4b66bdcb10\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00663957-2097-41ec-9a6e-3b4b66bdcb10\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f78f9ef6aa7daf827913b55deac0759d151d8207ef9542652868ea02faf95a5c/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.093436 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-284b5ec6-00ef-43bc-8818-098904e7931e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-284b5ec6-00ef-43bc-8818-098904e7931e\") pod \"logging-loki-compactor-0\" (UID: \"41d40d08-10d1-4abc-90e4-f894d3ccfa4b\") " pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.131325 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-00663957-2097-41ec-9a6e-3b4b66bdcb10\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00663957-2097-41ec-9a6e-3b4b66bdcb10\") pod \"logging-loki-index-gateway-0\" (UID: \"c00667ae-ca81-4c00-9ab2-71a1f307978d\") " pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.296517 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.389882 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.460665 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh"] Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.464853 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff"] Oct 02 09:50:41 crc kubenswrapper[4771]: W1002 09:50:41.482568 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21daf834_8ac7_4137_933e_b7d3db27e225.slice/crio-55df87b9fe4a7703e96782fb6ad4295a6fa93fdbce21c8ca17c4d1d0fbe09b7b WatchSource:0}: Error finding container 55df87b9fe4a7703e96782fb6ad4295a6fa93fdbce21c8ca17c4d1d0fbe09b7b: Status 404 returned error can't find the container with id 55df87b9fe4a7703e96782fb6ad4295a6fa93fdbce21c8ca17c4d1d0fbe09b7b Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.560019 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.587951 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.682699 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"34f9375e-2699-41ae-a604-81599cc16ea3","Type":"ContainerStarted","Data":"329f2bd59c8f49f2c4b06d5dcff5f28dc6021d0b4fa6e59e3bdd831f48aa4433"} Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.685100 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" event={"ID":"21daf834-8ac7-4137-933e-b7d3db27e225","Type":"ContainerStarted","Data":"55df87b9fe4a7703e96782fb6ad4295a6fa93fdbce21c8ca17c4d1d0fbe09b7b"} Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.686064 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" event={"ID":"cdb0a9ad-c8af-4298-9726-fa21864ee2a6","Type":"ContainerStarted","Data":"b4334c3c0eb06a09067ac1c2dcf72d347fc4c05addd1c1afce4eccf39d946f0b"} Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.686911 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" event={"ID":"f9c8afd6-efbc-4787-89ed-100438b03ecd","Type":"ContainerStarted","Data":"80e505d900e683e080c4251e72b9dc2b2489da604c1549f204116ee06a54b9c3"} Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.688375 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ssks" event={"ID":"173aa1ff-ba83-466e-b59c-fec9264277e4","Type":"ContainerStarted","Data":"bba547e6c276bc7db14acd3994a94fafac796d40c7546910916867961cbb99c8"} Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.696181 4771 generic.go:334] "Generic (PLEG): container finished" podID="98618258-d18c-4bf1-8456-0985f90d893a" containerID="903af8012c4e396f0e6215e9941b260ae0d151a4e56b01a6862ffc42c6038576" exitCode=0 Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.696372 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vm66q" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.720981 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6ssks" podStartSLOduration=2.270747897 podStartE2EDuration="5.720955604s" podCreationTimestamp="2025-10-02 09:50:36 +0000 UTC" firstStartedPulling="2025-10-02 09:50:37.629184822 +0000 UTC m=+825.276869879" lastFinishedPulling="2025-10-02 09:50:41.079392519 +0000 UTC m=+828.727077586" observedRunningTime="2025-10-02 09:50:41.716431872 +0000 UTC m=+829.364116939" watchObservedRunningTime="2025-10-02 09:50:41.720955604 +0000 UTC m=+829.368640691" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.739442 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98618258-d18c-4bf1-8456-0985f90d893a-catalog-content\") pod \"98618258-d18c-4bf1-8456-0985f90d893a\" (UID: \"98618258-d18c-4bf1-8456-0985f90d893a\") " Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.739526 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98618258-d18c-4bf1-8456-0985f90d893a-utilities\") pod \"98618258-d18c-4bf1-8456-0985f90d893a\" (UID: \"98618258-d18c-4bf1-8456-0985f90d893a\") " Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.739655 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blh64\" (UniqueName: \"kubernetes.io/projected/98618258-d18c-4bf1-8456-0985f90d893a-kube-api-access-blh64\") pod \"98618258-d18c-4bf1-8456-0985f90d893a\" (UID: \"98618258-d18c-4bf1-8456-0985f90d893a\") " Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.740704 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98618258-d18c-4bf1-8456-0985f90d893a-utilities" (OuterVolumeSpecName: "utilities") pod "98618258-d18c-4bf1-8456-0985f90d893a" (UID: "98618258-d18c-4bf1-8456-0985f90d893a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.747081 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm66q" event={"ID":"98618258-d18c-4bf1-8456-0985f90d893a","Type":"ContainerDied","Data":"903af8012c4e396f0e6215e9941b260ae0d151a4e56b01a6862ffc42c6038576"} Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.747157 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm66q" event={"ID":"98618258-d18c-4bf1-8456-0985f90d893a","Type":"ContainerDied","Data":"87ed52152199c6c6ad2c322159cd1906dce087b16c1173fc55651e9cfd6ad932"} Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.747185 4771 scope.go:117] "RemoveContainer" containerID="903af8012c4e396f0e6215e9941b260ae0d151a4e56b01a6862ffc42c6038576" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.748019 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98618258-d18c-4bf1-8456-0985f90d893a-kube-api-access-blh64" (OuterVolumeSpecName: "kube-api-access-blh64") pod "98618258-d18c-4bf1-8456-0985f90d893a" (UID: "98618258-d18c-4bf1-8456-0985f90d893a"). InnerVolumeSpecName "kube-api-access-blh64". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.778338 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.788732 4771 scope.go:117] "RemoveContainer" containerID="58ac9ced0014d7e1200fb5688c78f4c7226a9ec8d28574a4429e6e48504316ec" Oct 02 09:50:41 crc kubenswrapper[4771]: W1002 09:50:41.796321 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41d40d08_10d1_4abc_90e4_f894d3ccfa4b.slice/crio-3ee025b3dc2cf587e13950ed8b4600336672f560cbd73b61c269fd20fc209a3a WatchSource:0}: Error finding container 3ee025b3dc2cf587e13950ed8b4600336672f560cbd73b61c269fd20fc209a3a: Status 404 returned error can't find the container with id 3ee025b3dc2cf587e13950ed8b4600336672f560cbd73b61c269fd20fc209a3a Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.804661 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98618258-d18c-4bf1-8456-0985f90d893a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98618258-d18c-4bf1-8456-0985f90d893a" (UID: "98618258-d18c-4bf1-8456-0985f90d893a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.841303 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98618258-d18c-4bf1-8456-0985f90d893a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.841347 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98618258-d18c-4bf1-8456-0985f90d893a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.841357 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blh64\" (UniqueName: \"kubernetes.io/projected/98618258-d18c-4bf1-8456-0985f90d893a-kube-api-access-blh64\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.850353 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.854266 4771 scope.go:117] "RemoveContainer" containerID="cfa00909d241c35eab96270ae3b2021ec0ed489d238c9eaa1efd83d5522ab2bd" Oct 02 09:50:41 crc kubenswrapper[4771]: W1002 09:50:41.860996 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc00667ae_ca81_4c00_9ab2_71a1f307978d.slice/crio-d453fe89f700cfc7f6360c37a6f681fdda8071a2480c5d35c9472fdcb0871839 WatchSource:0}: Error finding container d453fe89f700cfc7f6360c37a6f681fdda8071a2480c5d35c9472fdcb0871839: Status 404 returned error can't find the container with id d453fe89f700cfc7f6360c37a6f681fdda8071a2480c5d35c9472fdcb0871839 Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.901788 4771 scope.go:117] "RemoveContainer" containerID="903af8012c4e396f0e6215e9941b260ae0d151a4e56b01a6862ffc42c6038576" Oct 02 09:50:41 crc kubenswrapper[4771]: E1002 09:50:41.902634 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"903af8012c4e396f0e6215e9941b260ae0d151a4e56b01a6862ffc42c6038576\": container with ID starting with 903af8012c4e396f0e6215e9941b260ae0d151a4e56b01a6862ffc42c6038576 not found: ID does not exist" containerID="903af8012c4e396f0e6215e9941b260ae0d151a4e56b01a6862ffc42c6038576" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.902686 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"903af8012c4e396f0e6215e9941b260ae0d151a4e56b01a6862ffc42c6038576"} err="failed to get container status \"903af8012c4e396f0e6215e9941b260ae0d151a4e56b01a6862ffc42c6038576\": rpc error: code = NotFound desc = could not find container \"903af8012c4e396f0e6215e9941b260ae0d151a4e56b01a6862ffc42c6038576\": container with ID starting with 903af8012c4e396f0e6215e9941b260ae0d151a4e56b01a6862ffc42c6038576 not found: ID does not exist" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.902715 4771 scope.go:117] "RemoveContainer" containerID="58ac9ced0014d7e1200fb5688c78f4c7226a9ec8d28574a4429e6e48504316ec" Oct 02 09:50:41 crc kubenswrapper[4771]: E1002 09:50:41.903234 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58ac9ced0014d7e1200fb5688c78f4c7226a9ec8d28574a4429e6e48504316ec\": container with ID starting with 58ac9ced0014d7e1200fb5688c78f4c7226a9ec8d28574a4429e6e48504316ec not found: ID does not exist" containerID="58ac9ced0014d7e1200fb5688c78f4c7226a9ec8d28574a4429e6e48504316ec" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.903271 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58ac9ced0014d7e1200fb5688c78f4c7226a9ec8d28574a4429e6e48504316ec"} err="failed to get container status \"58ac9ced0014d7e1200fb5688c78f4c7226a9ec8d28574a4429e6e48504316ec\": rpc error: code = NotFound desc = could not find container \"58ac9ced0014d7e1200fb5688c78f4c7226a9ec8d28574a4429e6e48504316ec\": container with ID starting with 58ac9ced0014d7e1200fb5688c78f4c7226a9ec8d28574a4429e6e48504316ec not found: ID does not exist" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.903287 4771 scope.go:117] "RemoveContainer" containerID="cfa00909d241c35eab96270ae3b2021ec0ed489d238c9eaa1efd83d5522ab2bd" Oct 02 09:50:41 crc kubenswrapper[4771]: E1002 09:50:41.903641 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfa00909d241c35eab96270ae3b2021ec0ed489d238c9eaa1efd83d5522ab2bd\": container with ID starting with cfa00909d241c35eab96270ae3b2021ec0ed489d238c9eaa1efd83d5522ab2bd not found: ID does not exist" containerID="cfa00909d241c35eab96270ae3b2021ec0ed489d238c9eaa1efd83d5522ab2bd" Oct 02 09:50:41 crc kubenswrapper[4771]: I1002 09:50:41.903665 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfa00909d241c35eab96270ae3b2021ec0ed489d238c9eaa1efd83d5522ab2bd"} err="failed to get container status \"cfa00909d241c35eab96270ae3b2021ec0ed489d238c9eaa1efd83d5522ab2bd\": rpc error: code = NotFound desc = could not find container \"cfa00909d241c35eab96270ae3b2021ec0ed489d238c9eaa1efd83d5522ab2bd\": container with ID starting with cfa00909d241c35eab96270ae3b2021ec0ed489d238c9eaa1efd83d5522ab2bd not found: ID does not exist" Oct 02 09:50:42 crc kubenswrapper[4771]: I1002 09:50:42.054354 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vm66q"] Oct 02 09:50:42 crc kubenswrapper[4771]: I1002 09:50:42.063781 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vm66q"] Oct 02 09:50:42 crc kubenswrapper[4771]: I1002 09:50:42.705539 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"41d40d08-10d1-4abc-90e4-f894d3ccfa4b","Type":"ContainerStarted","Data":"3ee025b3dc2cf587e13950ed8b4600336672f560cbd73b61c269fd20fc209a3a"} Oct 02 09:50:42 crc kubenswrapper[4771]: I1002 09:50:42.707071 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"c00667ae-ca81-4c00-9ab2-71a1f307978d","Type":"ContainerStarted","Data":"d453fe89f700cfc7f6360c37a6f681fdda8071a2480c5d35c9472fdcb0871839"} Oct 02 09:50:43 crc kubenswrapper[4771]: I1002 09:50:43.703779 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98618258-d18c-4bf1-8456-0985f90d893a" path="/var/lib/kubelet/pods/98618258-d18c-4bf1-8456-0985f90d893a/volumes" Oct 02 09:50:46 crc kubenswrapper[4771]: I1002 09:50:46.722872 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:46 crc kubenswrapper[4771]: I1002 09:50:46.722994 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:46 crc kubenswrapper[4771]: I1002 09:50:46.756381 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" event={"ID":"cdb0a9ad-c8af-4298-9726-fa21864ee2a6","Type":"ContainerStarted","Data":"006f5a5b5ab2fc8610c1d9d2ac26a6e31b2b18494a0f00989ded50c1fd231d7b"} Oct 02 09:50:46 crc kubenswrapper[4771]: I1002 09:50:46.757974 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" event={"ID":"f9c8afd6-efbc-4787-89ed-100438b03ecd","Type":"ContainerStarted","Data":"d33fc2bebf616df0e38a55e7d8d27a25c4ac83f64e2cf06ce2bf7175212cdfc2"} Oct 02 09:50:46 crc kubenswrapper[4771]: I1002 09:50:46.759626 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"34f9375e-2699-41ae-a604-81599cc16ea3","Type":"ContainerStarted","Data":"d9342687fd65bd1870622de65926c1167c7d3e8f8a9943e3e925251858479b9a"} Oct 02 09:50:46 crc kubenswrapper[4771]: I1002 09:50:46.761114 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" event={"ID":"21daf834-8ac7-4137-933e-b7d3db27e225","Type":"ContainerStarted","Data":"c4bc20dd36441a520971fb974bed108b96fefbc8c054fd227ba3a8f429dcd014"} Oct 02 09:50:46 crc kubenswrapper[4771]: I1002 09:50:46.770175 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" event={"ID":"e9014359-f9dc-4e80-83c1-127149a3a60e","Type":"ContainerStarted","Data":"50101d60e77c241a30051a218d997eac4aaa7b3acc2c02b4250230852ee60191"} Oct 02 09:50:46 crc kubenswrapper[4771]: I1002 09:50:46.771842 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"41d40d08-10d1-4abc-90e4-f894d3ccfa4b","Type":"ContainerStarted","Data":"15a9c964998f56d2e65f57efa141b245411805823a5dd5d503468ef651885bd5"} Oct 02 09:50:46 crc kubenswrapper[4771]: I1002 09:50:46.773536 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" event={"ID":"aca1f6a0-c677-4129-84ec-0d33b51193d2","Type":"ContainerStarted","Data":"8054156adc21e84e6cd0ef54baf613a77464faee25b43048503245249fe2b306"} Oct 02 09:50:46 crc kubenswrapper[4771]: I1002 09:50:46.850645 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:46 crc kubenswrapper[4771]: I1002 09:50:46.897227 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:47 crc kubenswrapper[4771]: I1002 09:50:47.185566 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6ssks"] Oct 02 09:50:47 crc kubenswrapper[4771]: I1002 09:50:47.780257 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"c00667ae-ca81-4c00-9ab2-71a1f307978d","Type":"ContainerStarted","Data":"9b0ae64faf7900da3499bf0147c5a196c7105cc6be4e24a256dd468791779330"} Oct 02 09:50:48 crc kubenswrapper[4771]: I1002 09:50:48.787752 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:50:48 crc kubenswrapper[4771]: I1002 09:50:48.787849 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:50:48 crc kubenswrapper[4771]: I1002 09:50:48.788185 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6ssks" podUID="173aa1ff-ba83-466e-b59c-fec9264277e4" containerName="registry-server" containerID="cri-o://bba547e6c276bc7db14acd3994a94fafac796d40c7546910916867961cbb99c8" gracePeriod=2 Oct 02 09:50:48 crc kubenswrapper[4771]: I1002 09:50:48.788413 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:50:48 crc kubenswrapper[4771]: I1002 09:50:48.811222 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=5.645891721 podStartE2EDuration="9.811206808s" podCreationTimestamp="2025-10-02 09:50:39 +0000 UTC" firstStartedPulling="2025-10-02 09:50:41.79857678 +0000 UTC m=+829.446261847" lastFinishedPulling="2025-10-02 09:50:45.963891867 +0000 UTC m=+833.611576934" observedRunningTime="2025-10-02 09:50:48.810924111 +0000 UTC m=+836.458609168" watchObservedRunningTime="2025-10-02 09:50:48.811206808 +0000 UTC m=+836.458891875" Oct 02 09:50:48 crc kubenswrapper[4771]: I1002 09:50:48.842371 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" podStartSLOduration=4.297338384 podStartE2EDuration="9.842350223s" podCreationTimestamp="2025-10-02 09:50:39 +0000 UTC" firstStartedPulling="2025-10-02 09:50:40.399598315 +0000 UTC m=+828.047283392" lastFinishedPulling="2025-10-02 09:50:45.944610164 +0000 UTC m=+833.592295231" observedRunningTime="2025-10-02 09:50:48.832213284 +0000 UTC m=+836.479898361" watchObservedRunningTime="2025-10-02 09:50:48.842350223 +0000 UTC m=+836.490035290" Oct 02 09:50:48 crc kubenswrapper[4771]: I1002 09:50:48.859187 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" podStartSLOduration=4.338993798 podStartE2EDuration="9.859169626s" podCreationTimestamp="2025-10-02 09:50:39 +0000 UTC" firstStartedPulling="2025-10-02 09:50:40.424383084 +0000 UTC m=+828.072068151" lastFinishedPulling="2025-10-02 09:50:45.944558912 +0000 UTC m=+833.592243979" observedRunningTime="2025-10-02 09:50:48.858472639 +0000 UTC m=+836.506157706" watchObservedRunningTime="2025-10-02 09:50:48.859169626 +0000 UTC m=+836.506854693" Oct 02 09:50:48 crc kubenswrapper[4771]: I1002 09:50:48.890206 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" podStartSLOduration=4.760077229 podStartE2EDuration="9.890181928s" podCreationTimestamp="2025-10-02 09:50:39 +0000 UTC" firstStartedPulling="2025-10-02 09:50:40.75120437 +0000 UTC m=+828.398889437" lastFinishedPulling="2025-10-02 09:50:45.881309069 +0000 UTC m=+833.528994136" observedRunningTime="2025-10-02 09:50:48.880166242 +0000 UTC m=+836.527851309" watchObservedRunningTime="2025-10-02 09:50:48.890181928 +0000 UTC m=+836.537866995" Oct 02 09:50:48 crc kubenswrapper[4771]: I1002 09:50:48.902645 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=5.7879973 podStartE2EDuration="9.902623873s" podCreationTimestamp="2025-10-02 09:50:39 +0000 UTC" firstStartedPulling="2025-10-02 09:50:41.865100673 +0000 UTC m=+829.512785750" lastFinishedPulling="2025-10-02 09:50:45.979727256 +0000 UTC m=+833.627412323" observedRunningTime="2025-10-02 09:50:48.899410684 +0000 UTC m=+836.547095781" watchObservedRunningTime="2025-10-02 09:50:48.902623873 +0000 UTC m=+836.550308940" Oct 02 09:50:48 crc kubenswrapper[4771]: I1002 09:50:48.933962 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=5.5727500039999995 podStartE2EDuration="9.933938642s" podCreationTimestamp="2025-10-02 09:50:39 +0000 UTC" firstStartedPulling="2025-10-02 09:50:41.583133129 +0000 UTC m=+829.230818196" lastFinishedPulling="2025-10-02 09:50:45.944321767 +0000 UTC m=+833.592006834" observedRunningTime="2025-10-02 09:50:48.929412811 +0000 UTC m=+836.577097878" watchObservedRunningTime="2025-10-02 09:50:48.933938642 +0000 UTC m=+836.581623709" Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.252272 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.378856 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173aa1ff-ba83-466e-b59c-fec9264277e4-catalog-content\") pod \"173aa1ff-ba83-466e-b59c-fec9264277e4\" (UID: \"173aa1ff-ba83-466e-b59c-fec9264277e4\") " Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.378963 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173aa1ff-ba83-466e-b59c-fec9264277e4-utilities\") pod \"173aa1ff-ba83-466e-b59c-fec9264277e4\" (UID: \"173aa1ff-ba83-466e-b59c-fec9264277e4\") " Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.379152 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjbpq\" (UniqueName: \"kubernetes.io/projected/173aa1ff-ba83-466e-b59c-fec9264277e4-kube-api-access-sjbpq\") pod \"173aa1ff-ba83-466e-b59c-fec9264277e4\" (UID: \"173aa1ff-ba83-466e-b59c-fec9264277e4\") " Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.380511 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/173aa1ff-ba83-466e-b59c-fec9264277e4-utilities" (OuterVolumeSpecName: "utilities") pod "173aa1ff-ba83-466e-b59c-fec9264277e4" (UID: "173aa1ff-ba83-466e-b59c-fec9264277e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.380670 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/173aa1ff-ba83-466e-b59c-fec9264277e4-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.387614 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/173aa1ff-ba83-466e-b59c-fec9264277e4-kube-api-access-sjbpq" (OuterVolumeSpecName: "kube-api-access-sjbpq") pod "173aa1ff-ba83-466e-b59c-fec9264277e4" (UID: "173aa1ff-ba83-466e-b59c-fec9264277e4"). InnerVolumeSpecName "kube-api-access-sjbpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.482116 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjbpq\" (UniqueName: \"kubernetes.io/projected/173aa1ff-ba83-466e-b59c-fec9264277e4-kube-api-access-sjbpq\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.709756 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/173aa1ff-ba83-466e-b59c-fec9264277e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "173aa1ff-ba83-466e-b59c-fec9264277e4" (UID: "173aa1ff-ba83-466e-b59c-fec9264277e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.796978 4771 generic.go:334] "Generic (PLEG): container finished" podID="173aa1ff-ba83-466e-b59c-fec9264277e4" containerID="bba547e6c276bc7db14acd3994a94fafac796d40c7546910916867961cbb99c8" exitCode=0 Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.798027 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ssks" Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.798476 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ssks" event={"ID":"173aa1ff-ba83-466e-b59c-fec9264277e4","Type":"ContainerDied","Data":"bba547e6c276bc7db14acd3994a94fafac796d40c7546910916867961cbb99c8"} Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.798503 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ssks" event={"ID":"173aa1ff-ba83-466e-b59c-fec9264277e4","Type":"ContainerDied","Data":"55ca1a4e15e18a496e6f1001bd93f60ff44b0959ca2ea48f11ff7b93428ffd2c"} Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.798518 4771 scope.go:117] "RemoveContainer" containerID="bba547e6c276bc7db14acd3994a94fafac796d40c7546910916867961cbb99c8" Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.800935 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/173aa1ff-ba83-466e-b59c-fec9264277e4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.820518 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6ssks"] Oct 02 09:50:49 crc kubenswrapper[4771]: I1002 09:50:49.827030 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6ssks"] Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.126962 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.296826 4771 scope.go:117] "RemoveContainer" containerID="ebffc7b604c96964683ccd5f3fea873620c02a044d237f51ede834a95f9b1a6b" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.394818 4771 scope.go:117] "RemoveContainer" containerID="a7de4d8e13dc4ce9746c347f4a46369846bc8794a45adf616c69686efadd0d60" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.429819 4771 scope.go:117] "RemoveContainer" containerID="bba547e6c276bc7db14acd3994a94fafac796d40c7546910916867961cbb99c8" Oct 02 09:50:50 crc kubenswrapper[4771]: E1002 09:50:50.430865 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bba547e6c276bc7db14acd3994a94fafac796d40c7546910916867961cbb99c8\": container with ID starting with bba547e6c276bc7db14acd3994a94fafac796d40c7546910916867961cbb99c8 not found: ID does not exist" containerID="bba547e6c276bc7db14acd3994a94fafac796d40c7546910916867961cbb99c8" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.430955 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bba547e6c276bc7db14acd3994a94fafac796d40c7546910916867961cbb99c8"} err="failed to get container status \"bba547e6c276bc7db14acd3994a94fafac796d40c7546910916867961cbb99c8\": rpc error: code = NotFound desc = could not find container \"bba547e6c276bc7db14acd3994a94fafac796d40c7546910916867961cbb99c8\": container with ID starting with bba547e6c276bc7db14acd3994a94fafac796d40c7546910916867961cbb99c8 not found: ID does not exist" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.431007 4771 scope.go:117] "RemoveContainer" containerID="ebffc7b604c96964683ccd5f3fea873620c02a044d237f51ede834a95f9b1a6b" Oct 02 09:50:50 crc kubenswrapper[4771]: E1002 09:50:50.431482 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebffc7b604c96964683ccd5f3fea873620c02a044d237f51ede834a95f9b1a6b\": container with ID starting with ebffc7b604c96964683ccd5f3fea873620c02a044d237f51ede834a95f9b1a6b not found: ID does not exist" containerID="ebffc7b604c96964683ccd5f3fea873620c02a044d237f51ede834a95f9b1a6b" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.431535 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebffc7b604c96964683ccd5f3fea873620c02a044d237f51ede834a95f9b1a6b"} err="failed to get container status \"ebffc7b604c96964683ccd5f3fea873620c02a044d237f51ede834a95f9b1a6b\": rpc error: code = NotFound desc = could not find container \"ebffc7b604c96964683ccd5f3fea873620c02a044d237f51ede834a95f9b1a6b\": container with ID starting with ebffc7b604c96964683ccd5f3fea873620c02a044d237f51ede834a95f9b1a6b not found: ID does not exist" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.431567 4771 scope.go:117] "RemoveContainer" containerID="a7de4d8e13dc4ce9746c347f4a46369846bc8794a45adf616c69686efadd0d60" Oct 02 09:50:50 crc kubenswrapper[4771]: E1002 09:50:50.432000 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7de4d8e13dc4ce9746c347f4a46369846bc8794a45adf616c69686efadd0d60\": container with ID starting with a7de4d8e13dc4ce9746c347f4a46369846bc8794a45adf616c69686efadd0d60 not found: ID does not exist" containerID="a7de4d8e13dc4ce9746c347f4a46369846bc8794a45adf616c69686efadd0d60" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.432071 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7de4d8e13dc4ce9746c347f4a46369846bc8794a45adf616c69686efadd0d60"} err="failed to get container status \"a7de4d8e13dc4ce9746c347f4a46369846bc8794a45adf616c69686efadd0d60\": rpc error: code = NotFound desc = could not find container \"a7de4d8e13dc4ce9746c347f4a46369846bc8794a45adf616c69686efadd0d60\": container with ID starting with a7de4d8e13dc4ce9746c347f4a46369846bc8794a45adf616c69686efadd0d60 not found: ID does not exist" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.805819 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" event={"ID":"f9c8afd6-efbc-4787-89ed-100438b03ecd","Type":"ContainerStarted","Data":"c1d1354ea9d3ca4dd3559fae5f209cf200278b71b1d9d125113a84f5ce6412d2"} Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.806003 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.806050 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.810060 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" event={"ID":"21daf834-8ac7-4137-933e-b7d3db27e225","Type":"ContainerStarted","Data":"10e2bfa72fb2a2adb403b860debb9fd6455b7f07e833a7381a30adfa5e1bd891"} Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.810320 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.810459 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.818257 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.818306 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.826300 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.827925 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-vvdff" podStartSLOduration=2.91670469 podStartE2EDuration="11.827912583s" podCreationTimestamp="2025-10-02 09:50:39 +0000 UTC" firstStartedPulling="2025-10-02 09:50:41.536837602 +0000 UTC m=+829.184522669" lastFinishedPulling="2025-10-02 09:50:50.448045495 +0000 UTC m=+838.095730562" observedRunningTime="2025-10-02 09:50:50.82617467 +0000 UTC m=+838.473859737" watchObservedRunningTime="2025-10-02 09:50:50.827912583 +0000 UTC m=+838.475597650" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.837682 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.867958 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-6dc97c5fb-2nswh" podStartSLOduration=2.958582079 podStartE2EDuration="11.867935846s" podCreationTimestamp="2025-10-02 09:50:39 +0000 UTC" firstStartedPulling="2025-10-02 09:50:41.531999223 +0000 UTC m=+829.179684290" lastFinishedPulling="2025-10-02 09:50:50.44135299 +0000 UTC m=+838.089038057" observedRunningTime="2025-10-02 09:50:50.867836894 +0000 UTC m=+838.515521961" watchObservedRunningTime="2025-10-02 09:50:50.867935846 +0000 UTC m=+838.515620913" Oct 02 09:50:50 crc kubenswrapper[4771]: I1002 09:50:50.916195 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:50:51 crc kubenswrapper[4771]: I1002 09:50:51.390611 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:50:51 crc kubenswrapper[4771]: I1002 09:50:51.689641 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="173aa1ff-ba83-466e-b59c-fec9264277e4" path="/var/lib/kubelet/pods/173aa1ff-ba83-466e-b59c-fec9264277e4/volumes" Oct 02 09:51:01 crc kubenswrapper[4771]: I1002 09:51:01.396831 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Oct 02 09:51:09 crc kubenswrapper[4771]: I1002 09:51:09.757866 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-6f5f7fff97-47hsj" Oct 02 09:51:09 crc kubenswrapper[4771]: I1002 09:51:09.952772 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5d954896cf-74tjf" Oct 02 09:51:10 crc kubenswrapper[4771]: I1002 09:51:10.134932 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-6fbbbc8b7d-twxfl" Oct 02 09:51:10 crc kubenswrapper[4771]: I1002 09:51:10.922866 4771 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Oct 02 09:51:10 crc kubenswrapper[4771]: I1002 09:51:10.922924 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="34f9375e-2699-41ae-a604-81599cc16ea3" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 02 09:51:11 crc kubenswrapper[4771]: I1002 09:51:11.302608 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.184885 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w22nx"] Oct 02 09:51:12 crc kubenswrapper[4771]: E1002 09:51:12.185187 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98618258-d18c-4bf1-8456-0985f90d893a" containerName="registry-server" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.185198 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="98618258-d18c-4bf1-8456-0985f90d893a" containerName="registry-server" Oct 02 09:51:12 crc kubenswrapper[4771]: E1002 09:51:12.185208 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="173aa1ff-ba83-466e-b59c-fec9264277e4" containerName="extract-utilities" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.185214 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="173aa1ff-ba83-466e-b59c-fec9264277e4" containerName="extract-utilities" Oct 02 09:51:12 crc kubenswrapper[4771]: E1002 09:51:12.185228 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98618258-d18c-4bf1-8456-0985f90d893a" containerName="extract-content" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.185234 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="98618258-d18c-4bf1-8456-0985f90d893a" containerName="extract-content" Oct 02 09:51:12 crc kubenswrapper[4771]: E1002 09:51:12.185248 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="173aa1ff-ba83-466e-b59c-fec9264277e4" containerName="extract-content" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.185254 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="173aa1ff-ba83-466e-b59c-fec9264277e4" containerName="extract-content" Oct 02 09:51:12 crc kubenswrapper[4771]: E1002 09:51:12.185267 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98618258-d18c-4bf1-8456-0985f90d893a" containerName="extract-utilities" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.185273 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="98618258-d18c-4bf1-8456-0985f90d893a" containerName="extract-utilities" Oct 02 09:51:12 crc kubenswrapper[4771]: E1002 09:51:12.185283 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="173aa1ff-ba83-466e-b59c-fec9264277e4" containerName="registry-server" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.185290 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="173aa1ff-ba83-466e-b59c-fec9264277e4" containerName="registry-server" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.185430 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="98618258-d18c-4bf1-8456-0985f90d893a" containerName="registry-server" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.185441 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="173aa1ff-ba83-466e-b59c-fec9264277e4" containerName="registry-server" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.186459 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.197469 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w22nx"] Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.364059 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f612b42-2271-41a5-ab02-5abe0d50d017-utilities\") pod \"redhat-marketplace-w22nx\" (UID: \"6f612b42-2271-41a5-ab02-5abe0d50d017\") " pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.364229 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gjpm\" (UniqueName: \"kubernetes.io/projected/6f612b42-2271-41a5-ab02-5abe0d50d017-kube-api-access-8gjpm\") pod \"redhat-marketplace-w22nx\" (UID: \"6f612b42-2271-41a5-ab02-5abe0d50d017\") " pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.364278 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f612b42-2271-41a5-ab02-5abe0d50d017-catalog-content\") pod \"redhat-marketplace-w22nx\" (UID: \"6f612b42-2271-41a5-ab02-5abe0d50d017\") " pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.466326 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f612b42-2271-41a5-ab02-5abe0d50d017-utilities\") pod \"redhat-marketplace-w22nx\" (UID: \"6f612b42-2271-41a5-ab02-5abe0d50d017\") " pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.466476 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gjpm\" (UniqueName: \"kubernetes.io/projected/6f612b42-2271-41a5-ab02-5abe0d50d017-kube-api-access-8gjpm\") pod \"redhat-marketplace-w22nx\" (UID: \"6f612b42-2271-41a5-ab02-5abe0d50d017\") " pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.466513 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f612b42-2271-41a5-ab02-5abe0d50d017-catalog-content\") pod \"redhat-marketplace-w22nx\" (UID: \"6f612b42-2271-41a5-ab02-5abe0d50d017\") " pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.466969 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f612b42-2271-41a5-ab02-5abe0d50d017-utilities\") pod \"redhat-marketplace-w22nx\" (UID: \"6f612b42-2271-41a5-ab02-5abe0d50d017\") " pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.467001 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f612b42-2271-41a5-ab02-5abe0d50d017-catalog-content\") pod \"redhat-marketplace-w22nx\" (UID: \"6f612b42-2271-41a5-ab02-5abe0d50d017\") " pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.488816 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gjpm\" (UniqueName: \"kubernetes.io/projected/6f612b42-2271-41a5-ab02-5abe0d50d017-kube-api-access-8gjpm\") pod \"redhat-marketplace-w22nx\" (UID: \"6f612b42-2271-41a5-ab02-5abe0d50d017\") " pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.513327 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:12 crc kubenswrapper[4771]: I1002 09:51:12.980605 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w22nx"] Oct 02 09:51:13 crc kubenswrapper[4771]: I1002 09:51:13.971406 4771 generic.go:334] "Generic (PLEG): container finished" podID="6f612b42-2271-41a5-ab02-5abe0d50d017" containerID="95b0599cdd4e70ad11780b35fdc83e2292c0c58c6c1ed31dd812cfb11bf23c29" exitCode=0 Oct 02 09:51:13 crc kubenswrapper[4771]: I1002 09:51:13.971473 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w22nx" event={"ID":"6f612b42-2271-41a5-ab02-5abe0d50d017","Type":"ContainerDied","Data":"95b0599cdd4e70ad11780b35fdc83e2292c0c58c6c1ed31dd812cfb11bf23c29"} Oct 02 09:51:13 crc kubenswrapper[4771]: I1002 09:51:13.971735 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w22nx" event={"ID":"6f612b42-2271-41a5-ab02-5abe0d50d017","Type":"ContainerStarted","Data":"fe9185c31600700102e7d5c760b5f5dc3d1eaec341eb595426561ad08f6db578"} Oct 02 09:51:14 crc kubenswrapper[4771]: I1002 09:51:14.994057 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w22nx" event={"ID":"6f612b42-2271-41a5-ab02-5abe0d50d017","Type":"ContainerStarted","Data":"45457b40534e81bd132b0d7598a20795aafaaae1802fe7e1890740020a89cf72"} Oct 02 09:51:16 crc kubenswrapper[4771]: I1002 09:51:16.000955 4771 generic.go:334] "Generic (PLEG): container finished" podID="6f612b42-2271-41a5-ab02-5abe0d50d017" containerID="45457b40534e81bd132b0d7598a20795aafaaae1802fe7e1890740020a89cf72" exitCode=0 Oct 02 09:51:16 crc kubenswrapper[4771]: I1002 09:51:16.001012 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w22nx" event={"ID":"6f612b42-2271-41a5-ab02-5abe0d50d017","Type":"ContainerDied","Data":"45457b40534e81bd132b0d7598a20795aafaaae1802fe7e1890740020a89cf72"} Oct 02 09:51:17 crc kubenswrapper[4771]: I1002 09:51:17.010717 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w22nx" event={"ID":"6f612b42-2271-41a5-ab02-5abe0d50d017","Type":"ContainerStarted","Data":"8a42b0558f8baa63c675b5992c0d44638eab7b5a0134f48da01fa18a65c5a797"} Oct 02 09:51:17 crc kubenswrapper[4771]: I1002 09:51:17.031900 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w22nx" podStartSLOduration=2.54942796 podStartE2EDuration="5.031880322s" podCreationTimestamp="2025-10-02 09:51:12 +0000 UTC" firstStartedPulling="2025-10-02 09:51:13.97317139 +0000 UTC m=+861.620856457" lastFinishedPulling="2025-10-02 09:51:16.455623752 +0000 UTC m=+864.103308819" observedRunningTime="2025-10-02 09:51:17.026317996 +0000 UTC m=+864.674003063" watchObservedRunningTime="2025-10-02 09:51:17.031880322 +0000 UTC m=+864.679565389" Oct 02 09:51:20 crc kubenswrapper[4771]: I1002 09:51:20.947357 4771 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Oct 02 09:51:20 crc kubenswrapper[4771]: I1002 09:51:20.947996 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="34f9375e-2699-41ae-a604-81599cc16ea3" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 02 09:51:22 crc kubenswrapper[4771]: I1002 09:51:22.514184 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:22 crc kubenswrapper[4771]: I1002 09:51:22.514555 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:22 crc kubenswrapper[4771]: I1002 09:51:22.561983 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:23 crc kubenswrapper[4771]: I1002 09:51:23.101366 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:23 crc kubenswrapper[4771]: I1002 09:51:23.163934 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w22nx"] Oct 02 09:51:25 crc kubenswrapper[4771]: I1002 09:51:25.064904 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w22nx" podUID="6f612b42-2271-41a5-ab02-5abe0d50d017" containerName="registry-server" containerID="cri-o://8a42b0558f8baa63c675b5992c0d44638eab7b5a0134f48da01fa18a65c5a797" gracePeriod=2 Oct 02 09:51:25 crc kubenswrapper[4771]: I1002 09:51:25.459537 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:25 crc kubenswrapper[4771]: I1002 09:51:25.560582 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gjpm\" (UniqueName: \"kubernetes.io/projected/6f612b42-2271-41a5-ab02-5abe0d50d017-kube-api-access-8gjpm\") pod \"6f612b42-2271-41a5-ab02-5abe0d50d017\" (UID: \"6f612b42-2271-41a5-ab02-5abe0d50d017\") " Oct 02 09:51:25 crc kubenswrapper[4771]: I1002 09:51:25.560643 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f612b42-2271-41a5-ab02-5abe0d50d017-utilities\") pod \"6f612b42-2271-41a5-ab02-5abe0d50d017\" (UID: \"6f612b42-2271-41a5-ab02-5abe0d50d017\") " Oct 02 09:51:25 crc kubenswrapper[4771]: I1002 09:51:25.560677 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f612b42-2271-41a5-ab02-5abe0d50d017-catalog-content\") pod \"6f612b42-2271-41a5-ab02-5abe0d50d017\" (UID: \"6f612b42-2271-41a5-ab02-5abe0d50d017\") " Oct 02 09:51:25 crc kubenswrapper[4771]: I1002 09:51:25.561955 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f612b42-2271-41a5-ab02-5abe0d50d017-utilities" (OuterVolumeSpecName: "utilities") pod "6f612b42-2271-41a5-ab02-5abe0d50d017" (UID: "6f612b42-2271-41a5-ab02-5abe0d50d017"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:51:25 crc kubenswrapper[4771]: I1002 09:51:25.565983 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f612b42-2271-41a5-ab02-5abe0d50d017-kube-api-access-8gjpm" (OuterVolumeSpecName: "kube-api-access-8gjpm") pod "6f612b42-2271-41a5-ab02-5abe0d50d017" (UID: "6f612b42-2271-41a5-ab02-5abe0d50d017"). InnerVolumeSpecName "kube-api-access-8gjpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:51:25 crc kubenswrapper[4771]: I1002 09:51:25.574965 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f612b42-2271-41a5-ab02-5abe0d50d017-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6f612b42-2271-41a5-ab02-5abe0d50d017" (UID: "6f612b42-2271-41a5-ab02-5abe0d50d017"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:51:25 crc kubenswrapper[4771]: I1002 09:51:25.663173 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gjpm\" (UniqueName: \"kubernetes.io/projected/6f612b42-2271-41a5-ab02-5abe0d50d017-kube-api-access-8gjpm\") on node \"crc\" DevicePath \"\"" Oct 02 09:51:25 crc kubenswrapper[4771]: I1002 09:51:25.663205 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f612b42-2271-41a5-ab02-5abe0d50d017-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:51:25 crc kubenswrapper[4771]: I1002 09:51:25.663217 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f612b42-2271-41a5-ab02-5abe0d50d017-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.072742 4771 generic.go:334] "Generic (PLEG): container finished" podID="6f612b42-2271-41a5-ab02-5abe0d50d017" containerID="8a42b0558f8baa63c675b5992c0d44638eab7b5a0134f48da01fa18a65c5a797" exitCode=0 Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.072929 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w22nx" event={"ID":"6f612b42-2271-41a5-ab02-5abe0d50d017","Type":"ContainerDied","Data":"8a42b0558f8baa63c675b5992c0d44638eab7b5a0134f48da01fa18a65c5a797"} Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.073108 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w22nx" event={"ID":"6f612b42-2271-41a5-ab02-5abe0d50d017","Type":"ContainerDied","Data":"fe9185c31600700102e7d5c760b5f5dc3d1eaec341eb595426561ad08f6db578"} Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.073156 4771 scope.go:117] "RemoveContainer" containerID="8a42b0558f8baa63c675b5992c0d44638eab7b5a0134f48da01fa18a65c5a797" Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.073002 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w22nx" Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.093631 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w22nx"] Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.104606 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w22nx"] Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.108492 4771 scope.go:117] "RemoveContainer" containerID="45457b40534e81bd132b0d7598a20795aafaaae1802fe7e1890740020a89cf72" Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.125245 4771 scope.go:117] "RemoveContainer" containerID="95b0599cdd4e70ad11780b35fdc83e2292c0c58c6c1ed31dd812cfb11bf23c29" Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.150447 4771 scope.go:117] "RemoveContainer" containerID="8a42b0558f8baa63c675b5992c0d44638eab7b5a0134f48da01fa18a65c5a797" Oct 02 09:51:26 crc kubenswrapper[4771]: E1002 09:51:26.150952 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a42b0558f8baa63c675b5992c0d44638eab7b5a0134f48da01fa18a65c5a797\": container with ID starting with 8a42b0558f8baa63c675b5992c0d44638eab7b5a0134f48da01fa18a65c5a797 not found: ID does not exist" containerID="8a42b0558f8baa63c675b5992c0d44638eab7b5a0134f48da01fa18a65c5a797" Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.151010 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a42b0558f8baa63c675b5992c0d44638eab7b5a0134f48da01fa18a65c5a797"} err="failed to get container status \"8a42b0558f8baa63c675b5992c0d44638eab7b5a0134f48da01fa18a65c5a797\": rpc error: code = NotFound desc = could not find container \"8a42b0558f8baa63c675b5992c0d44638eab7b5a0134f48da01fa18a65c5a797\": container with ID starting with 8a42b0558f8baa63c675b5992c0d44638eab7b5a0134f48da01fa18a65c5a797 not found: ID does not exist" Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.151040 4771 scope.go:117] "RemoveContainer" containerID="45457b40534e81bd132b0d7598a20795aafaaae1802fe7e1890740020a89cf72" Oct 02 09:51:26 crc kubenswrapper[4771]: E1002 09:51:26.151438 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45457b40534e81bd132b0d7598a20795aafaaae1802fe7e1890740020a89cf72\": container with ID starting with 45457b40534e81bd132b0d7598a20795aafaaae1802fe7e1890740020a89cf72 not found: ID does not exist" containerID="45457b40534e81bd132b0d7598a20795aafaaae1802fe7e1890740020a89cf72" Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.151472 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45457b40534e81bd132b0d7598a20795aafaaae1802fe7e1890740020a89cf72"} err="failed to get container status \"45457b40534e81bd132b0d7598a20795aafaaae1802fe7e1890740020a89cf72\": rpc error: code = NotFound desc = could not find container \"45457b40534e81bd132b0d7598a20795aafaaae1802fe7e1890740020a89cf72\": container with ID starting with 45457b40534e81bd132b0d7598a20795aafaaae1802fe7e1890740020a89cf72 not found: ID does not exist" Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.151495 4771 scope.go:117] "RemoveContainer" containerID="95b0599cdd4e70ad11780b35fdc83e2292c0c58c6c1ed31dd812cfb11bf23c29" Oct 02 09:51:26 crc kubenswrapper[4771]: E1002 09:51:26.151737 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95b0599cdd4e70ad11780b35fdc83e2292c0c58c6c1ed31dd812cfb11bf23c29\": container with ID starting with 95b0599cdd4e70ad11780b35fdc83e2292c0c58c6c1ed31dd812cfb11bf23c29 not found: ID does not exist" containerID="95b0599cdd4e70ad11780b35fdc83e2292c0c58c6c1ed31dd812cfb11bf23c29" Oct 02 09:51:26 crc kubenswrapper[4771]: I1002 09:51:26.151778 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95b0599cdd4e70ad11780b35fdc83e2292c0c58c6c1ed31dd812cfb11bf23c29"} err="failed to get container status \"95b0599cdd4e70ad11780b35fdc83e2292c0c58c6c1ed31dd812cfb11bf23c29\": rpc error: code = NotFound desc = could not find container \"95b0599cdd4e70ad11780b35fdc83e2292c0c58c6c1ed31dd812cfb11bf23c29\": container with ID starting with 95b0599cdd4e70ad11780b35fdc83e2292c0c58c6c1ed31dd812cfb11bf23c29 not found: ID does not exist" Oct 02 09:51:27 crc kubenswrapper[4771]: I1002 09:51:27.729091 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f612b42-2271-41a5-ab02-5abe0d50d017" path="/var/lib/kubelet/pods/6f612b42-2271-41a5-ab02-5abe0d50d017/volumes" Oct 02 09:51:30 crc kubenswrapper[4771]: I1002 09:51:30.921433 4771 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Oct 02 09:51:30 crc kubenswrapper[4771]: I1002 09:51:30.921832 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="34f9375e-2699-41ae-a604-81599cc16ea3" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 02 09:51:40 crc kubenswrapper[4771]: I1002 09:51:40.948334 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Oct 02 09:51:42 crc kubenswrapper[4771]: I1002 09:51:42.145984 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:51:42 crc kubenswrapper[4771]: I1002 09:51:42.146054 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.589871 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-74zdp"] Oct 02 09:52:00 crc kubenswrapper[4771]: E1002 09:52:00.590781 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f612b42-2271-41a5-ab02-5abe0d50d017" containerName="extract-content" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.590803 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f612b42-2271-41a5-ab02-5abe0d50d017" containerName="extract-content" Oct 02 09:52:00 crc kubenswrapper[4771]: E1002 09:52:00.590821 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f612b42-2271-41a5-ab02-5abe0d50d017" containerName="registry-server" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.590829 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f612b42-2271-41a5-ab02-5abe0d50d017" containerName="registry-server" Oct 02 09:52:00 crc kubenswrapper[4771]: E1002 09:52:00.590847 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f612b42-2271-41a5-ab02-5abe0d50d017" containerName="extract-utilities" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.590858 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f612b42-2271-41a5-ab02-5abe0d50d017" containerName="extract-utilities" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.591058 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f612b42-2271-41a5-ab02-5abe0d50d017" containerName="registry-server" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.591778 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.595892 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.597192 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.597503 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.597883 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.606018 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-crxwz" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.610951 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-74zdp"] Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.612584 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.748233 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-trusted-ca\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.748289 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-datadir\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.748320 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-syslog-receiver\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.748364 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-entrypoint\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.748539 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-sa-token\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.748659 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-config\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.748745 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-tmp\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.748776 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-config-openshift-service-cacrt\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.748827 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-token\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.748914 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-metrics\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.748987 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bgh9\" (UniqueName: \"kubernetes.io/projected/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-kube-api-access-2bgh9\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.794869 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-74zdp"] Oct 02 09:52:00 crc kubenswrapper[4771]: E1002 09:52:00.795277 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-2bgh9 metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-74zdp" podUID="0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.850004 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-trusted-ca\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.850063 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-datadir\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.850092 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-syslog-receiver\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.850113 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-entrypoint\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.850172 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-datadir\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.850174 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-sa-token\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.850238 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-config\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.850265 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-tmp\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: E1002 09:52:00.850275 4771 secret.go:188] Couldn't get secret openshift-logging/collector-syslog-receiver: secret "collector-syslog-receiver" not found Oct 02 09:52:00 crc kubenswrapper[4771]: E1002 09:52:00.850391 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-syslog-receiver podName:0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41 nodeName:}" failed. No retries permitted until 2025-10-02 09:52:01.350338341 +0000 UTC m=+908.998023478 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "collector-syslog-receiver" (UniqueName: "kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-syslog-receiver") pod "collector-74zdp" (UID: "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41") : secret "collector-syslog-receiver" not found Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.850281 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-config-openshift-service-cacrt\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.850616 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-token\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.850688 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-metrics\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.850738 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bgh9\" (UniqueName: \"kubernetes.io/projected/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-kube-api-access-2bgh9\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.851192 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-config-openshift-service-cacrt\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.851289 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-entrypoint\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.851620 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-trusted-ca\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.851981 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-config\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.860837 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-token\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.867491 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-tmp\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.869920 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-metrics\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.870114 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-sa-token\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:00 crc kubenswrapper[4771]: I1002 09:52:00.878795 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bgh9\" (UniqueName: \"kubernetes.io/projected/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-kube-api-access-2bgh9\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.315186 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-74zdp" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.324841 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-74zdp" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.358102 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-syslog-receiver\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.361680 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-syslog-receiver\") pod \"collector-74zdp\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " pod="openshift-logging/collector-74zdp" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.459640 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-metrics\") pod \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.459708 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-config\") pod \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.459770 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-datadir\") pod \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.459827 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-tmp\") pod \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.459855 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bgh9\" (UniqueName: \"kubernetes.io/projected/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-kube-api-access-2bgh9\") pod \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.459896 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-syslog-receiver\") pod \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.459927 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-config-openshift-service-cacrt\") pod \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.459976 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-sa-token\") pod \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.460001 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-token\") pod \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.460040 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-trusted-ca\") pod \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.460065 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-entrypoint\") pod \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\" (UID: \"0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41\") " Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.460483 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-datadir" (OuterVolumeSpecName: "datadir") pod "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41" (UID: "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.460991 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41" (UID: "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.461039 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41" (UID: "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.462662 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41" (UID: "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.463012 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-sa-token" (OuterVolumeSpecName: "sa-token") pod "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41" (UID: "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.463373 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-metrics" (OuterVolumeSpecName: "metrics") pod "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41" (UID: "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.463982 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-token" (OuterVolumeSpecName: "collector-token") pod "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41" (UID: "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.464795 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-config" (OuterVolumeSpecName: "config") pod "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41" (UID: "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.464821 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41" (UID: "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.466915 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-tmp" (OuterVolumeSpecName: "tmp") pod "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41" (UID: "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.468170 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-kube-api-access-2bgh9" (OuterVolumeSpecName: "kube-api-access-2bgh9") pod "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41" (UID: "0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41"). InnerVolumeSpecName "kube-api-access-2bgh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.561697 4771 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.561746 4771 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.561760 4771 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.561776 4771 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-collector-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.561788 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.561801 4771 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-entrypoint\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.561809 4771 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.561818 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.561829 4771 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-datadir\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.561838 4771 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-tmp\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:01 crc kubenswrapper[4771]: I1002 09:52:01.561848 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bgh9\" (UniqueName: \"kubernetes.io/projected/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41-kube-api-access-2bgh9\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.321105 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-74zdp" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.366525 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-74zdp"] Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.375626 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-74zdp"] Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.383828 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-llprk"] Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.384889 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.387477 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.388695 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.388746 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-crxwz" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.388904 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.388986 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.389360 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-llprk"] Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.397587 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.481803 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/0d5fb871-be70-4492-b6c9-6f7b6684b890-entrypoint\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.481958 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d5fb871-be70-4492-b6c9-6f7b6684b890-trusted-ca\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.482001 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0d5fb871-be70-4492-b6c9-6f7b6684b890-tmp\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.482210 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/0d5fb871-be70-4492-b6c9-6f7b6684b890-config-openshift-service-cacrt\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.482344 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/0d5fb871-be70-4492-b6c9-6f7b6684b890-sa-token\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.482415 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/0d5fb871-be70-4492-b6c9-6f7b6684b890-collector-token\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.482457 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0d5fb871-be70-4492-b6c9-6f7b6684b890-collector-syslog-receiver\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.482554 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/0d5fb871-be70-4492-b6c9-6f7b6684b890-datadir\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.482710 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/0d5fb871-be70-4492-b6c9-6f7b6684b890-metrics\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.482792 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgnnm\" (UniqueName: \"kubernetes.io/projected/0d5fb871-be70-4492-b6c9-6f7b6684b890-kube-api-access-bgnnm\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.482838 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5fb871-be70-4492-b6c9-6f7b6684b890-config\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.584030 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/0d5fb871-be70-4492-b6c9-6f7b6684b890-sa-token\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.584121 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/0d5fb871-be70-4492-b6c9-6f7b6684b890-collector-token\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.584182 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0d5fb871-be70-4492-b6c9-6f7b6684b890-collector-syslog-receiver\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.584200 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/0d5fb871-be70-4492-b6c9-6f7b6684b890-datadir\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.584230 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/0d5fb871-be70-4492-b6c9-6f7b6684b890-metrics\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.584249 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgnnm\" (UniqueName: \"kubernetes.io/projected/0d5fb871-be70-4492-b6c9-6f7b6684b890-kube-api-access-bgnnm\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.584264 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5fb871-be70-4492-b6c9-6f7b6684b890-config\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.584294 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/0d5fb871-be70-4492-b6c9-6f7b6684b890-entrypoint\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.584330 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d5fb871-be70-4492-b6c9-6f7b6684b890-trusted-ca\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.584352 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0d5fb871-be70-4492-b6c9-6f7b6684b890-tmp\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.584380 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/0d5fb871-be70-4492-b6c9-6f7b6684b890-config-openshift-service-cacrt\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.585196 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/0d5fb871-be70-4492-b6c9-6f7b6684b890-datadir\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.586927 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/0d5fb871-be70-4492-b6c9-6f7b6684b890-entrypoint\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.587359 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5fb871-be70-4492-b6c9-6f7b6684b890-config\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.587731 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0d5fb871-be70-4492-b6c9-6f7b6684b890-trusted-ca\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.587840 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/0d5fb871-be70-4492-b6c9-6f7b6684b890-config-openshift-service-cacrt\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.589624 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0d5fb871-be70-4492-b6c9-6f7b6684b890-collector-syslog-receiver\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.589910 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/0d5fb871-be70-4492-b6c9-6f7b6684b890-metrics\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.591084 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/0d5fb871-be70-4492-b6c9-6f7b6684b890-collector-token\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.598048 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0d5fb871-be70-4492-b6c9-6f7b6684b890-tmp\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.602549 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgnnm\" (UniqueName: \"kubernetes.io/projected/0d5fb871-be70-4492-b6c9-6f7b6684b890-kube-api-access-bgnnm\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.602672 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/0d5fb871-be70-4492-b6c9-6f7b6684b890-sa-token\") pod \"collector-llprk\" (UID: \"0d5fb871-be70-4492-b6c9-6f7b6684b890\") " pod="openshift-logging/collector-llprk" Oct 02 09:52:02 crc kubenswrapper[4771]: I1002 09:52:02.701987 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-llprk" Oct 02 09:52:03 crc kubenswrapper[4771]: I1002 09:52:03.158090 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-llprk"] Oct 02 09:52:03 crc kubenswrapper[4771]: I1002 09:52:03.328497 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-llprk" event={"ID":"0d5fb871-be70-4492-b6c9-6f7b6684b890","Type":"ContainerStarted","Data":"be44fc1dc3514227a0c29473a534b7178bf0ad2a1f6bff331ae7c6d237e713d7"} Oct 02 09:52:03 crc kubenswrapper[4771]: I1002 09:52:03.694092 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41" path="/var/lib/kubelet/pods/0d5bd00e-7ce9-42f0-b6e3-df20d1ed1b41/volumes" Oct 02 09:52:11 crc kubenswrapper[4771]: I1002 09:52:11.395948 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-llprk" event={"ID":"0d5fb871-be70-4492-b6c9-6f7b6684b890","Type":"ContainerStarted","Data":"e70c9ea283dacae4719750e25003110aea42ed3ba72d1225a07790dfeb88e6e0"} Oct 02 09:52:11 crc kubenswrapper[4771]: I1002 09:52:11.423741 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-llprk" podStartSLOduration=1.483392216 podStartE2EDuration="9.42372068s" podCreationTimestamp="2025-10-02 09:52:02 +0000 UTC" firstStartedPulling="2025-10-02 09:52:03.168874675 +0000 UTC m=+910.816559752" lastFinishedPulling="2025-10-02 09:52:11.109203149 +0000 UTC m=+918.756888216" observedRunningTime="2025-10-02 09:52:11.416016682 +0000 UTC m=+919.063701749" watchObservedRunningTime="2025-10-02 09:52:11.42372068 +0000 UTC m=+919.071405747" Oct 02 09:52:12 crc kubenswrapper[4771]: I1002 09:52:12.146655 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:52:12 crc kubenswrapper[4771]: I1002 09:52:12.146994 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.146902 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.147601 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.148146 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.149051 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e881357965830ad6d553dbf5d22ee7a11da1ee52ace8a1671edcdf44ed33436b"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.149188 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://e881357965830ad6d553dbf5d22ee7a11da1ee52ace8a1671edcdf44ed33436b" gracePeriod=600 Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.657469 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="e881357965830ad6d553dbf5d22ee7a11da1ee52ace8a1671edcdf44ed33436b" exitCode=0 Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.657536 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"e881357965830ad6d553dbf5d22ee7a11da1ee52ace8a1671edcdf44ed33436b"} Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.657772 4771 scope.go:117] "RemoveContainer" containerID="8f7959ad99e22e4304788ce28b1466fbafe93b57e3295c1b849950830aa10434" Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.781686 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt"] Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.783085 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.785161 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.795717 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt"] Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.898190 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5vpf\" (UniqueName: \"kubernetes.io/projected/23531133-454e-4938-bcad-744b9ee79a26-kube-api-access-k5vpf\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt\" (UID: \"23531133-454e-4938-bcad-744b9ee79a26\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.898293 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23531133-454e-4938-bcad-744b9ee79a26-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt\" (UID: \"23531133-454e-4938-bcad-744b9ee79a26\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.898411 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23531133-454e-4938-bcad-744b9ee79a26-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt\" (UID: \"23531133-454e-4938-bcad-744b9ee79a26\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" Oct 02 09:52:42 crc kubenswrapper[4771]: I1002 09:52:42.999417 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23531133-454e-4938-bcad-744b9ee79a26-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt\" (UID: \"23531133-454e-4938-bcad-744b9ee79a26\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" Oct 02 09:52:43 crc kubenswrapper[4771]: I1002 09:52:43.000090 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5vpf\" (UniqueName: \"kubernetes.io/projected/23531133-454e-4938-bcad-744b9ee79a26-kube-api-access-k5vpf\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt\" (UID: \"23531133-454e-4938-bcad-744b9ee79a26\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" Oct 02 09:52:43 crc kubenswrapper[4771]: I1002 09:52:43.000219 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23531133-454e-4938-bcad-744b9ee79a26-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt\" (UID: \"23531133-454e-4938-bcad-744b9ee79a26\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" Oct 02 09:52:43 crc kubenswrapper[4771]: I1002 09:52:43.000352 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23531133-454e-4938-bcad-744b9ee79a26-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt\" (UID: \"23531133-454e-4938-bcad-744b9ee79a26\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" Oct 02 09:52:43 crc kubenswrapper[4771]: I1002 09:52:43.000681 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23531133-454e-4938-bcad-744b9ee79a26-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt\" (UID: \"23531133-454e-4938-bcad-744b9ee79a26\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" Oct 02 09:52:43 crc kubenswrapper[4771]: I1002 09:52:43.021182 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5vpf\" (UniqueName: \"kubernetes.io/projected/23531133-454e-4938-bcad-744b9ee79a26-kube-api-access-k5vpf\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt\" (UID: \"23531133-454e-4938-bcad-744b9ee79a26\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" Oct 02 09:52:43 crc kubenswrapper[4771]: I1002 09:52:43.101698 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" Oct 02 09:52:43 crc kubenswrapper[4771]: I1002 09:52:43.537820 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt"] Oct 02 09:52:43 crc kubenswrapper[4771]: I1002 09:52:43.666907 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" event={"ID":"23531133-454e-4938-bcad-744b9ee79a26","Type":"ContainerStarted","Data":"ed500534edae18eb576128f256e60c0bc785f9339021b367d4dda14e2c7ff316"} Oct 02 09:52:43 crc kubenswrapper[4771]: I1002 09:52:43.670269 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"9385c2320866e6d26f67de876609b5bcc62a83b76b282bedf5fb7cfa0fec4a42"} Oct 02 09:52:43 crc kubenswrapper[4771]: E1002 09:52:43.959997 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23531133_454e_4938_bcad_744b9ee79a26.slice/crio-f66ae037aab54a1c08a84fe20912813cbaca572549489b69a29a6a2223ac1b47.scope\": RecentStats: unable to find data in memory cache]" Oct 02 09:52:44 crc kubenswrapper[4771]: I1002 09:52:44.679419 4771 generic.go:334] "Generic (PLEG): container finished" podID="23531133-454e-4938-bcad-744b9ee79a26" containerID="f66ae037aab54a1c08a84fe20912813cbaca572549489b69a29a6a2223ac1b47" exitCode=0 Oct 02 09:52:44 crc kubenswrapper[4771]: I1002 09:52:44.679505 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" event={"ID":"23531133-454e-4938-bcad-744b9ee79a26","Type":"ContainerDied","Data":"f66ae037aab54a1c08a84fe20912813cbaca572549489b69a29a6a2223ac1b47"} Oct 02 09:52:47 crc kubenswrapper[4771]: I1002 09:52:47.716726 4771 generic.go:334] "Generic (PLEG): container finished" podID="23531133-454e-4938-bcad-744b9ee79a26" containerID="0f316b455598db2635541e6ae9973f1523c6557c4c3bf3938840998870a3de28" exitCode=0 Oct 02 09:52:47 crc kubenswrapper[4771]: I1002 09:52:47.717026 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" event={"ID":"23531133-454e-4938-bcad-744b9ee79a26","Type":"ContainerDied","Data":"0f316b455598db2635541e6ae9973f1523c6557c4c3bf3938840998870a3de28"} Oct 02 09:52:48 crc kubenswrapper[4771]: I1002 09:52:48.726794 4771 generic.go:334] "Generic (PLEG): container finished" podID="23531133-454e-4938-bcad-744b9ee79a26" containerID="c4693123a9ea55ee25ecd483e2f738a2c8ea98a589acc10d1a10ba39135236bd" exitCode=0 Oct 02 09:52:48 crc kubenswrapper[4771]: I1002 09:52:48.727010 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" event={"ID":"23531133-454e-4938-bcad-744b9ee79a26","Type":"ContainerDied","Data":"c4693123a9ea55ee25ecd483e2f738a2c8ea98a589acc10d1a10ba39135236bd"} Oct 02 09:52:50 crc kubenswrapper[4771]: I1002 09:52:50.023960 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" Oct 02 09:52:50 crc kubenswrapper[4771]: I1002 09:52:50.124295 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5vpf\" (UniqueName: \"kubernetes.io/projected/23531133-454e-4938-bcad-744b9ee79a26-kube-api-access-k5vpf\") pod \"23531133-454e-4938-bcad-744b9ee79a26\" (UID: \"23531133-454e-4938-bcad-744b9ee79a26\") " Oct 02 09:52:50 crc kubenswrapper[4771]: I1002 09:52:50.124350 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23531133-454e-4938-bcad-744b9ee79a26-util\") pod \"23531133-454e-4938-bcad-744b9ee79a26\" (UID: \"23531133-454e-4938-bcad-744b9ee79a26\") " Oct 02 09:52:50 crc kubenswrapper[4771]: I1002 09:52:50.124422 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23531133-454e-4938-bcad-744b9ee79a26-bundle\") pod \"23531133-454e-4938-bcad-744b9ee79a26\" (UID: \"23531133-454e-4938-bcad-744b9ee79a26\") " Oct 02 09:52:50 crc kubenswrapper[4771]: I1002 09:52:50.125606 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23531133-454e-4938-bcad-744b9ee79a26-bundle" (OuterVolumeSpecName: "bundle") pod "23531133-454e-4938-bcad-744b9ee79a26" (UID: "23531133-454e-4938-bcad-744b9ee79a26"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:50 crc kubenswrapper[4771]: I1002 09:52:50.135867 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23531133-454e-4938-bcad-744b9ee79a26-util" (OuterVolumeSpecName: "util") pod "23531133-454e-4938-bcad-744b9ee79a26" (UID: "23531133-454e-4938-bcad-744b9ee79a26"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:50 crc kubenswrapper[4771]: I1002 09:52:50.137375 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23531133-454e-4938-bcad-744b9ee79a26-kube-api-access-k5vpf" (OuterVolumeSpecName: "kube-api-access-k5vpf") pod "23531133-454e-4938-bcad-744b9ee79a26" (UID: "23531133-454e-4938-bcad-744b9ee79a26"). InnerVolumeSpecName "kube-api-access-k5vpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:50 crc kubenswrapper[4771]: I1002 09:52:50.226686 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5vpf\" (UniqueName: \"kubernetes.io/projected/23531133-454e-4938-bcad-744b9ee79a26-kube-api-access-k5vpf\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:50 crc kubenswrapper[4771]: I1002 09:52:50.226733 4771 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23531133-454e-4938-bcad-744b9ee79a26-util\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:50 crc kubenswrapper[4771]: I1002 09:52:50.226747 4771 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23531133-454e-4938-bcad-744b9ee79a26-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:50 crc kubenswrapper[4771]: I1002 09:52:50.747950 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" event={"ID":"23531133-454e-4938-bcad-744b9ee79a26","Type":"ContainerDied","Data":"ed500534edae18eb576128f256e60c0bc785f9339021b367d4dda14e2c7ff316"} Oct 02 09:52:50 crc kubenswrapper[4771]: I1002 09:52:50.747990 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed500534edae18eb576128f256e60c0bc785f9339021b367d4dda14e2c7ff316" Oct 02 09:52:50 crc kubenswrapper[4771]: I1002 09:52:50.748034 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt" Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.526256 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-lwrtk"] Oct 02 09:52:54 crc kubenswrapper[4771]: E1002 09:52:54.527694 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23531133-454e-4938-bcad-744b9ee79a26" containerName="pull" Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.527717 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="23531133-454e-4938-bcad-744b9ee79a26" containerName="pull" Oct 02 09:52:54 crc kubenswrapper[4771]: E1002 09:52:54.527744 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23531133-454e-4938-bcad-744b9ee79a26" containerName="extract" Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.527756 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="23531133-454e-4938-bcad-744b9ee79a26" containerName="extract" Oct 02 09:52:54 crc kubenswrapper[4771]: E1002 09:52:54.527773 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23531133-454e-4938-bcad-744b9ee79a26" containerName="util" Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.527782 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="23531133-454e-4938-bcad-744b9ee79a26" containerName="util" Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.527952 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="23531133-454e-4938-bcad-744b9ee79a26" containerName="extract" Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.528876 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwrtk" Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.532666 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.534835 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.536608 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-j6cfd" Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.538258 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-lwrtk"] Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.597755 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4v58\" (UniqueName: \"kubernetes.io/projected/0534552f-6859-4588-89ef-b2e5a036102b-kube-api-access-t4v58\") pod \"nmstate-operator-858ddd8f98-lwrtk\" (UID: \"0534552f-6859-4588-89ef-b2e5a036102b\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwrtk" Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.699996 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4v58\" (UniqueName: \"kubernetes.io/projected/0534552f-6859-4588-89ef-b2e5a036102b-kube-api-access-t4v58\") pod \"nmstate-operator-858ddd8f98-lwrtk\" (UID: \"0534552f-6859-4588-89ef-b2e5a036102b\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwrtk" Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.724345 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4v58\" (UniqueName: \"kubernetes.io/projected/0534552f-6859-4588-89ef-b2e5a036102b-kube-api-access-t4v58\") pod \"nmstate-operator-858ddd8f98-lwrtk\" (UID: \"0534552f-6859-4588-89ef-b2e5a036102b\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwrtk" Oct 02 09:52:54 crc kubenswrapper[4771]: I1002 09:52:54.857978 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwrtk" Oct 02 09:52:55 crc kubenswrapper[4771]: I1002 09:52:55.158841 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-lwrtk"] Oct 02 09:52:55 crc kubenswrapper[4771]: I1002 09:52:55.781263 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwrtk" event={"ID":"0534552f-6859-4588-89ef-b2e5a036102b","Type":"ContainerStarted","Data":"48a9059a21697a6f259b10b6bcc2ba8b576e45e4dcc9bd4d7a85a9e19928f118"} Oct 02 09:53:00 crc kubenswrapper[4771]: I1002 09:53:00.820862 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwrtk" event={"ID":"0534552f-6859-4588-89ef-b2e5a036102b","Type":"ContainerStarted","Data":"1cf9a077e1d259be0c404a5f2e61c5e53e9451483b5759b4c53f1c988740955a"} Oct 02 09:53:00 crc kubenswrapper[4771]: I1002 09:53:00.847700 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwrtk" podStartSLOduration=2.345458477 podStartE2EDuration="6.847676877s" podCreationTimestamp="2025-10-02 09:52:54 +0000 UTC" firstStartedPulling="2025-10-02 09:52:55.176020752 +0000 UTC m=+962.823705819" lastFinishedPulling="2025-10-02 09:52:59.678239152 +0000 UTC m=+967.325924219" observedRunningTime="2025-10-02 09:53:00.845348307 +0000 UTC m=+968.493033384" watchObservedRunningTime="2025-10-02 09:53:00.847676877 +0000 UTC m=+968.495361944" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.421314 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-2cnlz"] Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.425725 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2cnlz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.428264 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-vgbbg" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.430236 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz"] Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.432053 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.447885 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.465182 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz"] Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.474519 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-2cnlz"] Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.489250 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-jxnvz"] Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.490803 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.512724 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbb5p\" (UniqueName: \"kubernetes.io/projected/95e5893b-430f-450e-8e67-44e0582a1583-kube-api-access-vbb5p\") pod \"nmstate-metrics-fdff9cb8d-2cnlz\" (UID: \"95e5893b-430f-450e-8e67-44e0582a1583\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2cnlz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.513164 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdmsr\" (UniqueName: \"kubernetes.io/projected/291a7779-68be-4198-8171-eab959730ce4-kube-api-access-gdmsr\") pod \"nmstate-webhook-6cdbc54649-9gwrz\" (UID: \"291a7779-68be-4198-8171-eab959730ce4\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.513654 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/291a7779-68be-4198-8171-eab959730ce4-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-9gwrz\" (UID: \"291a7779-68be-4198-8171-eab959730ce4\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.615155 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n822w\" (UniqueName: \"kubernetes.io/projected/4b03aaef-425d-4ad0-85a2-04b3416b30e4-kube-api-access-n822w\") pod \"nmstate-handler-jxnvz\" (UID: \"4b03aaef-425d-4ad0-85a2-04b3416b30e4\") " pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.615451 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4b03aaef-425d-4ad0-85a2-04b3416b30e4-ovs-socket\") pod \"nmstate-handler-jxnvz\" (UID: \"4b03aaef-425d-4ad0-85a2-04b3416b30e4\") " pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.615599 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4b03aaef-425d-4ad0-85a2-04b3416b30e4-dbus-socket\") pod \"nmstate-handler-jxnvz\" (UID: \"4b03aaef-425d-4ad0-85a2-04b3416b30e4\") " pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.615761 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4b03aaef-425d-4ad0-85a2-04b3416b30e4-nmstate-lock\") pod \"nmstate-handler-jxnvz\" (UID: \"4b03aaef-425d-4ad0-85a2-04b3416b30e4\") " pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.615881 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbb5p\" (UniqueName: \"kubernetes.io/projected/95e5893b-430f-450e-8e67-44e0582a1583-kube-api-access-vbb5p\") pod \"nmstate-metrics-fdff9cb8d-2cnlz\" (UID: \"95e5893b-430f-450e-8e67-44e0582a1583\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2cnlz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.616037 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdmsr\" (UniqueName: \"kubernetes.io/projected/291a7779-68be-4198-8171-eab959730ce4-kube-api-access-gdmsr\") pod \"nmstate-webhook-6cdbc54649-9gwrz\" (UID: \"291a7779-68be-4198-8171-eab959730ce4\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.616191 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/291a7779-68be-4198-8171-eab959730ce4-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-9gwrz\" (UID: \"291a7779-68be-4198-8171-eab959730ce4\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" Oct 02 09:53:04 crc kubenswrapper[4771]: E1002 09:53:04.616353 4771 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 02 09:53:04 crc kubenswrapper[4771]: E1002 09:53:04.616417 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/291a7779-68be-4198-8171-eab959730ce4-tls-key-pair podName:291a7779-68be-4198-8171-eab959730ce4 nodeName:}" failed. No retries permitted until 2025-10-02 09:53:05.116400397 +0000 UTC m=+972.764085464 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/291a7779-68be-4198-8171-eab959730ce4-tls-key-pair") pod "nmstate-webhook-6cdbc54649-9gwrz" (UID: "291a7779-68be-4198-8171-eab959730ce4") : secret "openshift-nmstate-webhook" not found Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.619606 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl"] Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.621093 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.629085 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-8gp5v" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.630027 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.630846 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.640695 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl"] Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.651771 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbb5p\" (UniqueName: \"kubernetes.io/projected/95e5893b-430f-450e-8e67-44e0582a1583-kube-api-access-vbb5p\") pod \"nmstate-metrics-fdff9cb8d-2cnlz\" (UID: \"95e5893b-430f-450e-8e67-44e0582a1583\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2cnlz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.666897 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdmsr\" (UniqueName: \"kubernetes.io/projected/291a7779-68be-4198-8171-eab959730ce4-kube-api-access-gdmsr\") pod \"nmstate-webhook-6cdbc54649-9gwrz\" (UID: \"291a7779-68be-4198-8171-eab959730ce4\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.717690 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n822w\" (UniqueName: \"kubernetes.io/projected/4b03aaef-425d-4ad0-85a2-04b3416b30e4-kube-api-access-n822w\") pod \"nmstate-handler-jxnvz\" (UID: \"4b03aaef-425d-4ad0-85a2-04b3416b30e4\") " pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.717766 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4b03aaef-425d-4ad0-85a2-04b3416b30e4-ovs-socket\") pod \"nmstate-handler-jxnvz\" (UID: \"4b03aaef-425d-4ad0-85a2-04b3416b30e4\") " pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.717814 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3de3b126-5868-413f-840f-f1502a0182e8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-nlhsl\" (UID: \"3de3b126-5868-413f-840f-f1502a0182e8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.717847 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4b03aaef-425d-4ad0-85a2-04b3416b30e4-dbus-socket\") pod \"nmstate-handler-jxnvz\" (UID: \"4b03aaef-425d-4ad0-85a2-04b3416b30e4\") " pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.717882 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5plr\" (UniqueName: \"kubernetes.io/projected/3de3b126-5868-413f-840f-f1502a0182e8-kube-api-access-t5plr\") pod \"nmstate-console-plugin-6b874cbd85-nlhsl\" (UID: \"3de3b126-5868-413f-840f-f1502a0182e8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.717944 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4b03aaef-425d-4ad0-85a2-04b3416b30e4-nmstate-lock\") pod \"nmstate-handler-jxnvz\" (UID: \"4b03aaef-425d-4ad0-85a2-04b3416b30e4\") " pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.718032 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3de3b126-5868-413f-840f-f1502a0182e8-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-nlhsl\" (UID: \"3de3b126-5868-413f-840f-f1502a0182e8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.718231 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4b03aaef-425d-4ad0-85a2-04b3416b30e4-ovs-socket\") pod \"nmstate-handler-jxnvz\" (UID: \"4b03aaef-425d-4ad0-85a2-04b3416b30e4\") " pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.718364 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4b03aaef-425d-4ad0-85a2-04b3416b30e4-nmstate-lock\") pod \"nmstate-handler-jxnvz\" (UID: \"4b03aaef-425d-4ad0-85a2-04b3416b30e4\") " pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.718485 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4b03aaef-425d-4ad0-85a2-04b3416b30e4-dbus-socket\") pod \"nmstate-handler-jxnvz\" (UID: \"4b03aaef-425d-4ad0-85a2-04b3416b30e4\") " pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.741544 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n822w\" (UniqueName: \"kubernetes.io/projected/4b03aaef-425d-4ad0-85a2-04b3416b30e4-kube-api-access-n822w\") pod \"nmstate-handler-jxnvz\" (UID: \"4b03aaef-425d-4ad0-85a2-04b3416b30e4\") " pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.754107 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2cnlz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.818725 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-569495bbfd-vb9kq"] Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.820937 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.821225 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.820177 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3de3b126-5868-413f-840f-f1502a0182e8-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-nlhsl\" (UID: \"3de3b126-5868-413f-840f-f1502a0182e8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.822660 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3de3b126-5868-413f-840f-f1502a0182e8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-nlhsl\" (UID: \"3de3b126-5868-413f-840f-f1502a0182e8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.822824 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5plr\" (UniqueName: \"kubernetes.io/projected/3de3b126-5868-413f-840f-f1502a0182e8-kube-api-access-t5plr\") pod \"nmstate-console-plugin-6b874cbd85-nlhsl\" (UID: \"3de3b126-5868-413f-840f-f1502a0182e8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.824583 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3de3b126-5868-413f-840f-f1502a0182e8-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-nlhsl\" (UID: \"3de3b126-5868-413f-840f-f1502a0182e8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.830407 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3de3b126-5868-413f-840f-f1502a0182e8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-nlhsl\" (UID: \"3de3b126-5868-413f-840f-f1502a0182e8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.852765 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-569495bbfd-vb9kq"] Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.863042 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5plr\" (UniqueName: \"kubernetes.io/projected/3de3b126-5868-413f-840f-f1502a0182e8-kube-api-access-t5plr\") pod \"nmstate-console-plugin-6b874cbd85-nlhsl\" (UID: \"3de3b126-5868-413f-840f-f1502a0182e8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.925248 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-oauth-serving-cert\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.925633 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-trusted-ca-bundle\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.925792 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-serving-cert\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.925892 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-config\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.926015 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-service-ca\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.926350 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgplb\" (UniqueName: \"kubernetes.io/projected/e5be9e93-f0cb-41e5-b232-6f02e4337d59-kube-api-access-bgplb\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.926483 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-oauth-config\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:04 crc kubenswrapper[4771]: I1002 09:53:04.945236 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.029039 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-oauth-serving-cert\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.029540 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-trusted-ca-bundle\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.029596 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-serving-cert\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.029624 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-config\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.029641 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-service-ca\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.029696 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgplb\" (UniqueName: \"kubernetes.io/projected/e5be9e93-f0cb-41e5-b232-6f02e4337d59-kube-api-access-bgplb\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.029731 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-oauth-config\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.030198 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-oauth-serving-cert\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.030886 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-config\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.032095 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-service-ca\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.036611 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-trusted-ca-bundle\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.039350 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-oauth-config\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.039443 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-serving-cert\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.074818 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgplb\" (UniqueName: \"kubernetes.io/projected/e5be9e93-f0cb-41e5-b232-6f02e4337d59-kube-api-access-bgplb\") pod \"console-569495bbfd-vb9kq\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.131082 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/291a7779-68be-4198-8171-eab959730ce4-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-9gwrz\" (UID: \"291a7779-68be-4198-8171-eab959730ce4\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.137351 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/291a7779-68be-4198-8171-eab959730ce4-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-9gwrz\" (UID: \"291a7779-68be-4198-8171-eab959730ce4\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.161846 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.364227 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.390632 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-2cnlz"] Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.562630 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-569495bbfd-vb9kq"] Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.634761 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl"] Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.865933 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-jxnvz" event={"ID":"4b03aaef-425d-4ad0-85a2-04b3416b30e4","Type":"ContainerStarted","Data":"b7f3e84d39ff8b337185a070d212911444367fbc62253ee9a8372f26270af76a"} Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.867403 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-569495bbfd-vb9kq" event={"ID":"e5be9e93-f0cb-41e5-b232-6f02e4337d59","Type":"ContainerStarted","Data":"0ad6c8f8a8cc4b33f9a2647b4b650bba50991ccc352407fa684e96de80d8cad6"} Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.867454 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-569495bbfd-vb9kq" event={"ID":"e5be9e93-f0cb-41e5-b232-6f02e4337d59","Type":"ContainerStarted","Data":"a6f2fa9bcb3240d7aadc4dffab955b869f2b3477a670200c221ae903a32c661c"} Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.868924 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" event={"ID":"3de3b126-5868-413f-840f-f1502a0182e8","Type":"ContainerStarted","Data":"94c39e48cb2d423d02a1aa7d8f485d094d9cb1552361a943907440bfa26aca8a"} Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.869938 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2cnlz" event={"ID":"95e5893b-430f-450e-8e67-44e0582a1583","Type":"ContainerStarted","Data":"91086c8ac893995828c25dc6fda665851f2add0121c8858c8e1af645e8ce857a"} Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.892520 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-569495bbfd-vb9kq" podStartSLOduration=1.892497483 podStartE2EDuration="1.892497483s" podCreationTimestamp="2025-10-02 09:53:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:53:05.88908934 +0000 UTC m=+973.536774417" watchObservedRunningTime="2025-10-02 09:53:05.892497483 +0000 UTC m=+973.540182550" Oct 02 09:53:05 crc kubenswrapper[4771]: I1002 09:53:05.925020 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz"] Oct 02 09:53:06 crc kubenswrapper[4771]: I1002 09:53:06.878332 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" event={"ID":"291a7779-68be-4198-8171-eab959730ce4","Type":"ContainerStarted","Data":"8017dcacd9e897c2a2b1234bb85d35726af75f36aa1a2c029d13715cf28d6cbe"} Oct 02 09:53:09 crc kubenswrapper[4771]: I1002 09:53:09.902822 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" event={"ID":"291a7779-68be-4198-8171-eab959730ce4","Type":"ContainerStarted","Data":"64e6cf84acce8f71a6ebcb616c6ede46362fee59d4a482ed013d5c32d0211372"} Oct 02 09:53:09 crc kubenswrapper[4771]: I1002 09:53:09.903649 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" Oct 02 09:53:09 crc kubenswrapper[4771]: I1002 09:53:09.907489 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" event={"ID":"3de3b126-5868-413f-840f-f1502a0182e8","Type":"ContainerStarted","Data":"a5b8f0544363d4c06ac8c20cacace804e0f623cd5bcc6e0cc60edc41c912c456"} Oct 02 09:53:09 crc kubenswrapper[4771]: I1002 09:53:09.909995 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2cnlz" event={"ID":"95e5893b-430f-450e-8e67-44e0582a1583","Type":"ContainerStarted","Data":"53cd455b5d73f3487a8d41223286531385caed2001dd0546539a8c03726ed21d"} Oct 02 09:53:09 crc kubenswrapper[4771]: I1002 09:53:09.911785 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-jxnvz" event={"ID":"4b03aaef-425d-4ad0-85a2-04b3416b30e4","Type":"ContainerStarted","Data":"990741bd34c2d6074107a4faa6cdebee910a6dcbc1456bd070b53bb12dfbc336"} Oct 02 09:53:09 crc kubenswrapper[4771]: I1002 09:53:09.911916 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:09 crc kubenswrapper[4771]: I1002 09:53:09.943631 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" podStartSLOduration=2.8036008 podStartE2EDuration="5.943606607s" podCreationTimestamp="2025-10-02 09:53:04 +0000 UTC" firstStartedPulling="2025-10-02 09:53:05.941389216 +0000 UTC m=+973.589074283" lastFinishedPulling="2025-10-02 09:53:09.081395023 +0000 UTC m=+976.729080090" observedRunningTime="2025-10-02 09:53:09.92280498 +0000 UTC m=+977.570490067" watchObservedRunningTime="2025-10-02 09:53:09.943606607 +0000 UTC m=+977.591291674" Oct 02 09:53:09 crc kubenswrapper[4771]: I1002 09:53:09.964877 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-nlhsl" podStartSLOduration=2.556447638 podStartE2EDuration="5.964843476s" podCreationTimestamp="2025-10-02 09:53:04 +0000 UTC" firstStartedPulling="2025-10-02 09:53:05.673619481 +0000 UTC m=+973.321304548" lastFinishedPulling="2025-10-02 09:53:09.082015319 +0000 UTC m=+976.729700386" observedRunningTime="2025-10-02 09:53:09.939647681 +0000 UTC m=+977.587332748" watchObservedRunningTime="2025-10-02 09:53:09.964843476 +0000 UTC m=+977.612528553" Oct 02 09:53:12 crc kubenswrapper[4771]: I1002 09:53:12.935211 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2cnlz" event={"ID":"95e5893b-430f-450e-8e67-44e0582a1583","Type":"ContainerStarted","Data":"125f2e09f915b62a7ef2daa7e5337e9e0b4cff416d8cc203b0f084f4a6103a19"} Oct 02 09:53:12 crc kubenswrapper[4771]: I1002 09:53:12.951006 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2cnlz" podStartSLOduration=1.779871934 podStartE2EDuration="8.950987458s" podCreationTimestamp="2025-10-02 09:53:04 +0000 UTC" firstStartedPulling="2025-10-02 09:53:05.415359257 +0000 UTC m=+973.063044324" lastFinishedPulling="2025-10-02 09:53:12.586474781 +0000 UTC m=+980.234159848" observedRunningTime="2025-10-02 09:53:12.950703331 +0000 UTC m=+980.598388398" watchObservedRunningTime="2025-10-02 09:53:12.950987458 +0000 UTC m=+980.598672525" Oct 02 09:53:12 crc kubenswrapper[4771]: I1002 09:53:12.953509 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-jxnvz" podStartSLOduration=4.781077924 podStartE2EDuration="8.953500769s" podCreationTimestamp="2025-10-02 09:53:04 +0000 UTC" firstStartedPulling="2025-10-02 09:53:04.909910771 +0000 UTC m=+972.557595838" lastFinishedPulling="2025-10-02 09:53:09.082333606 +0000 UTC m=+976.730018683" observedRunningTime="2025-10-02 09:53:09.966205929 +0000 UTC m=+977.613891016" watchObservedRunningTime="2025-10-02 09:53:12.953500769 +0000 UTC m=+980.601185836" Oct 02 09:53:14 crc kubenswrapper[4771]: I1002 09:53:14.849940 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-jxnvz" Oct 02 09:53:15 crc kubenswrapper[4771]: I1002 09:53:15.163296 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:15 crc kubenswrapper[4771]: I1002 09:53:15.163452 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:15 crc kubenswrapper[4771]: I1002 09:53:15.168568 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:15 crc kubenswrapper[4771]: I1002 09:53:15.959875 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:53:16 crc kubenswrapper[4771]: I1002 09:53:16.032652 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-78bb69b5d5-cl8tx"] Oct 02 09:53:25 crc kubenswrapper[4771]: I1002 09:53:25.369765 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9gwrz" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.088685 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-78bb69b5d5-cl8tx" podUID="ff5471c7-c311-4bc5-984a-0793a3b53faf" containerName="console" containerID="cri-o://4417aaef2742c4ec4705b18db998c5d23ca7f126c03c2e21bcd198080fa98bbc" gracePeriod=15 Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.661311 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-78bb69b5d5-cl8tx_ff5471c7-c311-4bc5-984a-0793a3b53faf/console/0.log" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.661625 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.751866 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-service-ca\") pod \"ff5471c7-c311-4bc5-984a-0793a3b53faf\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.752258 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-serving-cert\") pod \"ff5471c7-c311-4bc5-984a-0793a3b53faf\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.752329 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-trusted-ca-bundle\") pod \"ff5471c7-c311-4bc5-984a-0793a3b53faf\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.752379 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-oauth-config\") pod \"ff5471c7-c311-4bc5-984a-0793a3b53faf\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.752401 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-oauth-serving-cert\") pod \"ff5471c7-c311-4bc5-984a-0793a3b53faf\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.752431 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhtwg\" (UniqueName: \"kubernetes.io/projected/ff5471c7-c311-4bc5-984a-0793a3b53faf-kube-api-access-mhtwg\") pod \"ff5471c7-c311-4bc5-984a-0793a3b53faf\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.752456 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-config\") pod \"ff5471c7-c311-4bc5-984a-0793a3b53faf\" (UID: \"ff5471c7-c311-4bc5-984a-0793a3b53faf\") " Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.753432 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-config" (OuterVolumeSpecName: "console-config") pod "ff5471c7-c311-4bc5-984a-0793a3b53faf" (UID: "ff5471c7-c311-4bc5-984a-0793a3b53faf"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.753652 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-service-ca" (OuterVolumeSpecName: "service-ca") pod "ff5471c7-c311-4bc5-984a-0793a3b53faf" (UID: "ff5471c7-c311-4bc5-984a-0793a3b53faf"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.753820 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "ff5471c7-c311-4bc5-984a-0793a3b53faf" (UID: "ff5471c7-c311-4bc5-984a-0793a3b53faf"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.754356 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "ff5471c7-c311-4bc5-984a-0793a3b53faf" (UID: "ff5471c7-c311-4bc5-984a-0793a3b53faf"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.758950 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "ff5471c7-c311-4bc5-984a-0793a3b53faf" (UID: "ff5471c7-c311-4bc5-984a-0793a3b53faf"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.759148 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff5471c7-c311-4bc5-984a-0793a3b53faf-kube-api-access-mhtwg" (OuterVolumeSpecName: "kube-api-access-mhtwg") pod "ff5471c7-c311-4bc5-984a-0793a3b53faf" (UID: "ff5471c7-c311-4bc5-984a-0793a3b53faf"). InnerVolumeSpecName "kube-api-access-mhtwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.765457 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "ff5471c7-c311-4bc5-984a-0793a3b53faf" (UID: "ff5471c7-c311-4bc5-984a-0793a3b53faf"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.854708 4771 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.854764 4771 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.854778 4771 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.854791 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.854809 4771 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ff5471c7-c311-4bc5-984a-0793a3b53faf-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.854821 4771 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ff5471c7-c311-4bc5-984a-0793a3b53faf-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:41 crc kubenswrapper[4771]: I1002 09:53:41.854834 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhtwg\" (UniqueName: \"kubernetes.io/projected/ff5471c7-c311-4bc5-984a-0793a3b53faf-kube-api-access-mhtwg\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:42 crc kubenswrapper[4771]: I1002 09:53:42.168761 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-78bb69b5d5-cl8tx_ff5471c7-c311-4bc5-984a-0793a3b53faf/console/0.log" Oct 02 09:53:42 crc kubenswrapper[4771]: I1002 09:53:42.168840 4771 generic.go:334] "Generic (PLEG): container finished" podID="ff5471c7-c311-4bc5-984a-0793a3b53faf" containerID="4417aaef2742c4ec4705b18db998c5d23ca7f126c03c2e21bcd198080fa98bbc" exitCode=2 Oct 02 09:53:42 crc kubenswrapper[4771]: I1002 09:53:42.168901 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-78bb69b5d5-cl8tx" event={"ID":"ff5471c7-c311-4bc5-984a-0793a3b53faf","Type":"ContainerDied","Data":"4417aaef2742c4ec4705b18db998c5d23ca7f126c03c2e21bcd198080fa98bbc"} Oct 02 09:53:42 crc kubenswrapper[4771]: I1002 09:53:42.168918 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-78bb69b5d5-cl8tx" Oct 02 09:53:42 crc kubenswrapper[4771]: I1002 09:53:42.168998 4771 scope.go:117] "RemoveContainer" containerID="4417aaef2742c4ec4705b18db998c5d23ca7f126c03c2e21bcd198080fa98bbc" Oct 02 09:53:42 crc kubenswrapper[4771]: I1002 09:53:42.168981 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-78bb69b5d5-cl8tx" event={"ID":"ff5471c7-c311-4bc5-984a-0793a3b53faf","Type":"ContainerDied","Data":"0a5dae2359324ea25f8e73e8587a23e5c803ef2ac78907ffc2d61be721871a76"} Oct 02 09:53:42 crc kubenswrapper[4771]: I1002 09:53:42.190011 4771 scope.go:117] "RemoveContainer" containerID="4417aaef2742c4ec4705b18db998c5d23ca7f126c03c2e21bcd198080fa98bbc" Oct 02 09:53:42 crc kubenswrapper[4771]: E1002 09:53:42.190796 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4417aaef2742c4ec4705b18db998c5d23ca7f126c03c2e21bcd198080fa98bbc\": container with ID starting with 4417aaef2742c4ec4705b18db998c5d23ca7f126c03c2e21bcd198080fa98bbc not found: ID does not exist" containerID="4417aaef2742c4ec4705b18db998c5d23ca7f126c03c2e21bcd198080fa98bbc" Oct 02 09:53:42 crc kubenswrapper[4771]: I1002 09:53:42.190866 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4417aaef2742c4ec4705b18db998c5d23ca7f126c03c2e21bcd198080fa98bbc"} err="failed to get container status \"4417aaef2742c4ec4705b18db998c5d23ca7f126c03c2e21bcd198080fa98bbc\": rpc error: code = NotFound desc = could not find container \"4417aaef2742c4ec4705b18db998c5d23ca7f126c03c2e21bcd198080fa98bbc\": container with ID starting with 4417aaef2742c4ec4705b18db998c5d23ca7f126c03c2e21bcd198080fa98bbc not found: ID does not exist" Oct 02 09:53:42 crc kubenswrapper[4771]: I1002 09:53:42.207747 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-78bb69b5d5-cl8tx"] Oct 02 09:53:42 crc kubenswrapper[4771]: I1002 09:53:42.215607 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-78bb69b5d5-cl8tx"] Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.052073 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5"] Oct 02 09:53:43 crc kubenswrapper[4771]: E1002 09:53:43.052786 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5471c7-c311-4bc5-984a-0793a3b53faf" containerName="console" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.052808 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5471c7-c311-4bc5-984a-0793a3b53faf" containerName="console" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.053056 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff5471c7-c311-4bc5-984a-0793a3b53faf" containerName="console" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.054404 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.058706 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.076271 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5"] Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.077319 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0e20c875-b90e-424e-9fe3-e873406a86db-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5\" (UID: \"0e20c875-b90e-424e-9fe3-e873406a86db\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.077482 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mn4q\" (UniqueName: \"kubernetes.io/projected/0e20c875-b90e-424e-9fe3-e873406a86db-kube-api-access-7mn4q\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5\" (UID: \"0e20c875-b90e-424e-9fe3-e873406a86db\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.077638 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0e20c875-b90e-424e-9fe3-e873406a86db-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5\" (UID: \"0e20c875-b90e-424e-9fe3-e873406a86db\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.179019 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0e20c875-b90e-424e-9fe3-e873406a86db-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5\" (UID: \"0e20c875-b90e-424e-9fe3-e873406a86db\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.179088 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mn4q\" (UniqueName: \"kubernetes.io/projected/0e20c875-b90e-424e-9fe3-e873406a86db-kube-api-access-7mn4q\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5\" (UID: \"0e20c875-b90e-424e-9fe3-e873406a86db\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.179179 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0e20c875-b90e-424e-9fe3-e873406a86db-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5\" (UID: \"0e20c875-b90e-424e-9fe3-e873406a86db\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.179698 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0e20c875-b90e-424e-9fe3-e873406a86db-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5\" (UID: \"0e20c875-b90e-424e-9fe3-e873406a86db\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.179777 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0e20c875-b90e-424e-9fe3-e873406a86db-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5\" (UID: \"0e20c875-b90e-424e-9fe3-e873406a86db\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.209368 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mn4q\" (UniqueName: \"kubernetes.io/projected/0e20c875-b90e-424e-9fe3-e873406a86db-kube-api-access-7mn4q\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5\" (UID: \"0e20c875-b90e-424e-9fe3-e873406a86db\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.369327 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.722546 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff5471c7-c311-4bc5-984a-0793a3b53faf" path="/var/lib/kubelet/pods/ff5471c7-c311-4bc5-984a-0793a3b53faf/volumes" Oct 02 09:53:43 crc kubenswrapper[4771]: I1002 09:53:43.838949 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5"] Oct 02 09:53:44 crc kubenswrapper[4771]: I1002 09:53:44.189662 4771 generic.go:334] "Generic (PLEG): container finished" podID="0e20c875-b90e-424e-9fe3-e873406a86db" containerID="de389cdb5fa762811f719c8813ae95c7ccc667d8afb6bd860b2ec9dba41ae2cc" exitCode=0 Oct 02 09:53:44 crc kubenswrapper[4771]: I1002 09:53:44.189727 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" event={"ID":"0e20c875-b90e-424e-9fe3-e873406a86db","Type":"ContainerDied","Data":"de389cdb5fa762811f719c8813ae95c7ccc667d8afb6bd860b2ec9dba41ae2cc"} Oct 02 09:53:44 crc kubenswrapper[4771]: I1002 09:53:44.189995 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" event={"ID":"0e20c875-b90e-424e-9fe3-e873406a86db","Type":"ContainerStarted","Data":"bae671f2c25d9ca1b0ad8a7b86a3e9f4b166ca11cd45843d18047f3da4e8f7d0"} Oct 02 09:53:44 crc kubenswrapper[4771]: I1002 09:53:44.191453 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:53:46 crc kubenswrapper[4771]: I1002 09:53:46.207935 4771 generic.go:334] "Generic (PLEG): container finished" podID="0e20c875-b90e-424e-9fe3-e873406a86db" containerID="46f3e94806457e1dcff99a877aa4de5f175764868e170119e0432da4522334ed" exitCode=0 Oct 02 09:53:46 crc kubenswrapper[4771]: I1002 09:53:46.208087 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" event={"ID":"0e20c875-b90e-424e-9fe3-e873406a86db","Type":"ContainerDied","Data":"46f3e94806457e1dcff99a877aa4de5f175764868e170119e0432da4522334ed"} Oct 02 09:53:47 crc kubenswrapper[4771]: I1002 09:53:47.219286 4771 generic.go:334] "Generic (PLEG): container finished" podID="0e20c875-b90e-424e-9fe3-e873406a86db" containerID="a53380c3d9cf1c31fdfaa97051753b3cec9f2fd3b4d40807add01983b1eb852f" exitCode=0 Oct 02 09:53:47 crc kubenswrapper[4771]: I1002 09:53:47.219391 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" event={"ID":"0e20c875-b90e-424e-9fe3-e873406a86db","Type":"ContainerDied","Data":"a53380c3d9cf1c31fdfaa97051753b3cec9f2fd3b4d40807add01983b1eb852f"} Oct 02 09:53:48 crc kubenswrapper[4771]: I1002 09:53:48.497454 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" Oct 02 09:53:48 crc kubenswrapper[4771]: I1002 09:53:48.581440 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0e20c875-b90e-424e-9fe3-e873406a86db-util\") pod \"0e20c875-b90e-424e-9fe3-e873406a86db\" (UID: \"0e20c875-b90e-424e-9fe3-e873406a86db\") " Oct 02 09:53:48 crc kubenswrapper[4771]: I1002 09:53:48.581651 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0e20c875-b90e-424e-9fe3-e873406a86db-bundle\") pod \"0e20c875-b90e-424e-9fe3-e873406a86db\" (UID: \"0e20c875-b90e-424e-9fe3-e873406a86db\") " Oct 02 09:53:48 crc kubenswrapper[4771]: I1002 09:53:48.581672 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mn4q\" (UniqueName: \"kubernetes.io/projected/0e20c875-b90e-424e-9fe3-e873406a86db-kube-api-access-7mn4q\") pod \"0e20c875-b90e-424e-9fe3-e873406a86db\" (UID: \"0e20c875-b90e-424e-9fe3-e873406a86db\") " Oct 02 09:53:48 crc kubenswrapper[4771]: I1002 09:53:48.582857 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e20c875-b90e-424e-9fe3-e873406a86db-bundle" (OuterVolumeSpecName: "bundle") pod "0e20c875-b90e-424e-9fe3-e873406a86db" (UID: "0e20c875-b90e-424e-9fe3-e873406a86db"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:53:48 crc kubenswrapper[4771]: I1002 09:53:48.591336 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e20c875-b90e-424e-9fe3-e873406a86db-kube-api-access-7mn4q" (OuterVolumeSpecName: "kube-api-access-7mn4q") pod "0e20c875-b90e-424e-9fe3-e873406a86db" (UID: "0e20c875-b90e-424e-9fe3-e873406a86db"). InnerVolumeSpecName "kube-api-access-7mn4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:53:48 crc kubenswrapper[4771]: I1002 09:53:48.611944 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e20c875-b90e-424e-9fe3-e873406a86db-util" (OuterVolumeSpecName: "util") pod "0e20c875-b90e-424e-9fe3-e873406a86db" (UID: "0e20c875-b90e-424e-9fe3-e873406a86db"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:53:48 crc kubenswrapper[4771]: I1002 09:53:48.683969 4771 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0e20c875-b90e-424e-9fe3-e873406a86db-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:48 crc kubenswrapper[4771]: I1002 09:53:48.684007 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mn4q\" (UniqueName: \"kubernetes.io/projected/0e20c875-b90e-424e-9fe3-e873406a86db-kube-api-access-7mn4q\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:48 crc kubenswrapper[4771]: I1002 09:53:48.684022 4771 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0e20c875-b90e-424e-9fe3-e873406a86db-util\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:49 crc kubenswrapper[4771]: I1002 09:53:49.235512 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" event={"ID":"0e20c875-b90e-424e-9fe3-e873406a86db","Type":"ContainerDied","Data":"bae671f2c25d9ca1b0ad8a7b86a3e9f4b166ca11cd45843d18047f3da4e8f7d0"} Oct 02 09:53:49 crc kubenswrapper[4771]: I1002 09:53:49.235902 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bae671f2c25d9ca1b0ad8a7b86a3e9f4b166ca11cd45843d18047f3da4e8f7d0" Oct 02 09:53:49 crc kubenswrapper[4771]: I1002 09:53:49.235605 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.887506 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck"] Oct 02 09:53:57 crc kubenswrapper[4771]: E1002 09:53:57.889331 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e20c875-b90e-424e-9fe3-e873406a86db" containerName="pull" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.889417 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e20c875-b90e-424e-9fe3-e873406a86db" containerName="pull" Oct 02 09:53:57 crc kubenswrapper[4771]: E1002 09:53:57.889519 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e20c875-b90e-424e-9fe3-e873406a86db" containerName="util" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.889591 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e20c875-b90e-424e-9fe3-e873406a86db" containerName="util" Oct 02 09:53:57 crc kubenswrapper[4771]: E1002 09:53:57.889671 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e20c875-b90e-424e-9fe3-e873406a86db" containerName="extract" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.889738 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e20c875-b90e-424e-9fe3-e873406a86db" containerName="extract" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.889946 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e20c875-b90e-424e-9fe3-e873406a86db" containerName="extract" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.890552 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.896422 4771 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-jmbh4" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.897603 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.897680 4771 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.897719 4771 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.898014 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.911435 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck"] Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.942146 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nvck\" (UniqueName: \"kubernetes.io/projected/6ff4fcd7-2222-4405-911d-6d337e1af8a2-kube-api-access-5nvck\") pod \"metallb-operator-controller-manager-7bc8798995-4p5ck\" (UID: \"6ff4fcd7-2222-4405-911d-6d337e1af8a2\") " pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.942236 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6ff4fcd7-2222-4405-911d-6d337e1af8a2-apiservice-cert\") pod \"metallb-operator-controller-manager-7bc8798995-4p5ck\" (UID: \"6ff4fcd7-2222-4405-911d-6d337e1af8a2\") " pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" Oct 02 09:53:57 crc kubenswrapper[4771]: I1002 09:53:57.942321 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6ff4fcd7-2222-4405-911d-6d337e1af8a2-webhook-cert\") pod \"metallb-operator-controller-manager-7bc8798995-4p5ck\" (UID: \"6ff4fcd7-2222-4405-911d-6d337e1af8a2\") " pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.044208 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6ff4fcd7-2222-4405-911d-6d337e1af8a2-webhook-cert\") pod \"metallb-operator-controller-manager-7bc8798995-4p5ck\" (UID: \"6ff4fcd7-2222-4405-911d-6d337e1af8a2\") " pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.045634 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nvck\" (UniqueName: \"kubernetes.io/projected/6ff4fcd7-2222-4405-911d-6d337e1af8a2-kube-api-access-5nvck\") pod \"metallb-operator-controller-manager-7bc8798995-4p5ck\" (UID: \"6ff4fcd7-2222-4405-911d-6d337e1af8a2\") " pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.045694 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6ff4fcd7-2222-4405-911d-6d337e1af8a2-apiservice-cert\") pod \"metallb-operator-controller-manager-7bc8798995-4p5ck\" (UID: \"6ff4fcd7-2222-4405-911d-6d337e1af8a2\") " pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.061097 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6ff4fcd7-2222-4405-911d-6d337e1af8a2-apiservice-cert\") pod \"metallb-operator-controller-manager-7bc8798995-4p5ck\" (UID: \"6ff4fcd7-2222-4405-911d-6d337e1af8a2\") " pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.075924 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nvck\" (UniqueName: \"kubernetes.io/projected/6ff4fcd7-2222-4405-911d-6d337e1af8a2-kube-api-access-5nvck\") pod \"metallb-operator-controller-manager-7bc8798995-4p5ck\" (UID: \"6ff4fcd7-2222-4405-911d-6d337e1af8a2\") " pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.082830 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6ff4fcd7-2222-4405-911d-6d337e1af8a2-webhook-cert\") pod \"metallb-operator-controller-manager-7bc8798995-4p5ck\" (UID: \"6ff4fcd7-2222-4405-911d-6d337e1af8a2\") " pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.212530 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.385343 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j"] Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.386681 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.389021 4771 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-lr9v8" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.389069 4771 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.389305 4771 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.405272 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j"] Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.453613 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbbk2\" (UniqueName: \"kubernetes.io/projected/8968c764-9017-43a2-bc9e-c2d3fe04a7fc-kube-api-access-sbbk2\") pod \"metallb-operator-webhook-server-8547c6bb6c-bjb6j\" (UID: \"8968c764-9017-43a2-bc9e-c2d3fe04a7fc\") " pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.453680 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8968c764-9017-43a2-bc9e-c2d3fe04a7fc-apiservice-cert\") pod \"metallb-operator-webhook-server-8547c6bb6c-bjb6j\" (UID: \"8968c764-9017-43a2-bc9e-c2d3fe04a7fc\") " pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.453749 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8968c764-9017-43a2-bc9e-c2d3fe04a7fc-webhook-cert\") pod \"metallb-operator-webhook-server-8547c6bb6c-bjb6j\" (UID: \"8968c764-9017-43a2-bc9e-c2d3fe04a7fc\") " pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.555601 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbbk2\" (UniqueName: \"kubernetes.io/projected/8968c764-9017-43a2-bc9e-c2d3fe04a7fc-kube-api-access-sbbk2\") pod \"metallb-operator-webhook-server-8547c6bb6c-bjb6j\" (UID: \"8968c764-9017-43a2-bc9e-c2d3fe04a7fc\") " pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.555655 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8968c764-9017-43a2-bc9e-c2d3fe04a7fc-apiservice-cert\") pod \"metallb-operator-webhook-server-8547c6bb6c-bjb6j\" (UID: \"8968c764-9017-43a2-bc9e-c2d3fe04a7fc\") " pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.555739 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8968c764-9017-43a2-bc9e-c2d3fe04a7fc-webhook-cert\") pod \"metallb-operator-webhook-server-8547c6bb6c-bjb6j\" (UID: \"8968c764-9017-43a2-bc9e-c2d3fe04a7fc\") " pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.561463 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8968c764-9017-43a2-bc9e-c2d3fe04a7fc-apiservice-cert\") pod \"metallb-operator-webhook-server-8547c6bb6c-bjb6j\" (UID: \"8968c764-9017-43a2-bc9e-c2d3fe04a7fc\") " pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.563639 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8968c764-9017-43a2-bc9e-c2d3fe04a7fc-webhook-cert\") pod \"metallb-operator-webhook-server-8547c6bb6c-bjb6j\" (UID: \"8968c764-9017-43a2-bc9e-c2d3fe04a7fc\") " pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.580167 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbbk2\" (UniqueName: \"kubernetes.io/projected/8968c764-9017-43a2-bc9e-c2d3fe04a7fc-kube-api-access-sbbk2\") pod \"metallb-operator-webhook-server-8547c6bb6c-bjb6j\" (UID: \"8968c764-9017-43a2-bc9e-c2d3fe04a7fc\") " pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.704102 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck"] Oct 02 09:53:58 crc kubenswrapper[4771]: W1002 09:53:58.709591 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff4fcd7_2222_4405_911d_6d337e1af8a2.slice/crio-caa221bfb0ad6d3d52a746e483dd6ef2286e8e2a40cce452bf3b02b87a740b1b WatchSource:0}: Error finding container caa221bfb0ad6d3d52a746e483dd6ef2286e8e2a40cce452bf3b02b87a740b1b: Status 404 returned error can't find the container with id caa221bfb0ad6d3d52a746e483dd6ef2286e8e2a40cce452bf3b02b87a740b1b Oct 02 09:53:58 crc kubenswrapper[4771]: I1002 09:53:58.725794 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" Oct 02 09:53:59 crc kubenswrapper[4771]: I1002 09:53:59.148993 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j"] Oct 02 09:53:59 crc kubenswrapper[4771]: W1002 09:53:59.154598 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8968c764_9017_43a2_bc9e_c2d3fe04a7fc.slice/crio-e38fcfb88d684418aaeabb6856f59bb7c16088f1bd71c01c06d7e050907ee83c WatchSource:0}: Error finding container e38fcfb88d684418aaeabb6856f59bb7c16088f1bd71c01c06d7e050907ee83c: Status 404 returned error can't find the container with id e38fcfb88d684418aaeabb6856f59bb7c16088f1bd71c01c06d7e050907ee83c Oct 02 09:53:59 crc kubenswrapper[4771]: I1002 09:53:59.317354 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" event={"ID":"6ff4fcd7-2222-4405-911d-6d337e1af8a2","Type":"ContainerStarted","Data":"caa221bfb0ad6d3d52a746e483dd6ef2286e8e2a40cce452bf3b02b87a740b1b"} Oct 02 09:53:59 crc kubenswrapper[4771]: I1002 09:53:59.319058 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" event={"ID":"8968c764-9017-43a2-bc9e-c2d3fe04a7fc","Type":"ContainerStarted","Data":"e38fcfb88d684418aaeabb6856f59bb7c16088f1bd71c01c06d7e050907ee83c"} Oct 02 09:54:04 crc kubenswrapper[4771]: I1002 09:54:04.368475 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" event={"ID":"8968c764-9017-43a2-bc9e-c2d3fe04a7fc","Type":"ContainerStarted","Data":"424dd565b7bed953cd987d855ff2e5d5d61d893d0ed574242bd27bd3bf04a806"} Oct 02 09:54:04 crc kubenswrapper[4771]: I1002 09:54:04.370071 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" Oct 02 09:54:04 crc kubenswrapper[4771]: I1002 09:54:04.371951 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" event={"ID":"6ff4fcd7-2222-4405-911d-6d337e1af8a2","Type":"ContainerStarted","Data":"df6b2f5cc0b40f78866ca7efdbdb306379387bf738decb1ce34abf1bf0b22380"} Oct 02 09:54:04 crc kubenswrapper[4771]: I1002 09:54:04.372488 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" Oct 02 09:54:04 crc kubenswrapper[4771]: I1002 09:54:04.407783 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" podStartSLOduration=1.5832097680000001 podStartE2EDuration="6.40776562s" podCreationTimestamp="2025-10-02 09:53:58 +0000 UTC" firstStartedPulling="2025-10-02 09:53:59.156182206 +0000 UTC m=+1026.803867273" lastFinishedPulling="2025-10-02 09:54:03.980738068 +0000 UTC m=+1031.628423125" observedRunningTime="2025-10-02 09:54:04.404750477 +0000 UTC m=+1032.052435544" watchObservedRunningTime="2025-10-02 09:54:04.40776562 +0000 UTC m=+1032.055450687" Oct 02 09:54:04 crc kubenswrapper[4771]: I1002 09:54:04.446014 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" podStartSLOduration=2.196480548 podStartE2EDuration="7.445996143s" podCreationTimestamp="2025-10-02 09:53:57 +0000 UTC" firstStartedPulling="2025-10-02 09:53:58.712812564 +0000 UTC m=+1026.360497631" lastFinishedPulling="2025-10-02 09:54:03.962328159 +0000 UTC m=+1031.610013226" observedRunningTime="2025-10-02 09:54:04.439327371 +0000 UTC m=+1032.087012438" watchObservedRunningTime="2025-10-02 09:54:04.445996143 +0000 UTC m=+1032.093681210" Oct 02 09:54:18 crc kubenswrapper[4771]: I1002 09:54:18.732438 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-8547c6bb6c-bjb6j" Oct 02 09:54:38 crc kubenswrapper[4771]: I1002 09:54:38.214943 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7bc8798995-4p5ck" Oct 02 09:54:38 crc kubenswrapper[4771]: I1002 09:54:38.996483 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-cf6h7"] Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.000699 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.002239 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4"] Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.003788 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.003953 4771 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.004845 4771 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-mpkx9" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.005005 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.015113 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4"] Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.018438 4771 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.083771 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5638abcc-fdb8-4ed6-a630-8562d79668a5-frr-sockets\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.083866 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3c832819-5b80-4548-8941-6e411bd1ba2f-cert\") pod \"frr-k8s-webhook-server-64bf5d555-qvlg4\" (UID: \"3c832819-5b80-4548-8941-6e411bd1ba2f\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.083918 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpnh2\" (UniqueName: \"kubernetes.io/projected/5638abcc-fdb8-4ed6-a630-8562d79668a5-kube-api-access-fpnh2\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.083954 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8bpv\" (UniqueName: \"kubernetes.io/projected/3c832819-5b80-4548-8941-6e411bd1ba2f-kube-api-access-c8bpv\") pod \"frr-k8s-webhook-server-64bf5d555-qvlg4\" (UID: \"3c832819-5b80-4548-8941-6e411bd1ba2f\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.083990 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5638abcc-fdb8-4ed6-a630-8562d79668a5-frr-startup\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.084032 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5638abcc-fdb8-4ed6-a630-8562d79668a5-frr-conf\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.084141 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5638abcc-fdb8-4ed6-a630-8562d79668a5-metrics\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.084191 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5638abcc-fdb8-4ed6-a630-8562d79668a5-reloader\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.084326 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5638abcc-fdb8-4ed6-a630-8562d79668a5-metrics-certs\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.134384 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-nrj2m"] Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.136015 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-nrj2m" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.143687 4771 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.144181 4771 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.146630 4771 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-gbldw" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.149472 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-hsbdv"] Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.151463 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.153353 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.156246 4771 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.180108 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-hsbdv"] Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.185547 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pl5d\" (UniqueName: \"kubernetes.io/projected/381993c1-33fd-485c-bb38-3f18cfa91abf-kube-api-access-7pl5d\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.185615 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3c832819-5b80-4548-8941-6e411bd1ba2f-cert\") pod \"frr-k8s-webhook-server-64bf5d555-qvlg4\" (UID: \"3c832819-5b80-4548-8941-6e411bd1ba2f\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.185657 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpnh2\" (UniqueName: \"kubernetes.io/projected/5638abcc-fdb8-4ed6-a630-8562d79668a5-kube-api-access-fpnh2\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.185698 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8bpv\" (UniqueName: \"kubernetes.io/projected/3c832819-5b80-4548-8941-6e411bd1ba2f-kube-api-access-c8bpv\") pod \"frr-k8s-webhook-server-64bf5d555-qvlg4\" (UID: \"3c832819-5b80-4548-8941-6e411bd1ba2f\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.185724 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5638abcc-fdb8-4ed6-a630-8562d79668a5-frr-startup\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.185771 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5638abcc-fdb8-4ed6-a630-8562d79668a5-frr-conf\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.185825 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/381993c1-33fd-485c-bb38-3f18cfa91abf-memberlist\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.185884 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5638abcc-fdb8-4ed6-a630-8562d79668a5-metrics\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.185937 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5638abcc-fdb8-4ed6-a630-8562d79668a5-reloader\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.185969 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpj2j\" (UniqueName: \"kubernetes.io/projected/37d57598-c98a-45ff-8de0-c51a5e39adeb-kube-api-access-tpj2j\") pod \"controller-68d546b9d8-hsbdv\" (UID: \"37d57598-c98a-45ff-8de0-c51a5e39adeb\") " pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.186054 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/37d57598-c98a-45ff-8de0-c51a5e39adeb-cert\") pod \"controller-68d546b9d8-hsbdv\" (UID: \"37d57598-c98a-45ff-8de0-c51a5e39adeb\") " pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.186167 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/381993c1-33fd-485c-bb38-3f18cfa91abf-metrics-certs\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.186250 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5638abcc-fdb8-4ed6-a630-8562d79668a5-metrics-certs\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.186318 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/381993c1-33fd-485c-bb38-3f18cfa91abf-metallb-excludel2\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.186365 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37d57598-c98a-45ff-8de0-c51a5e39adeb-metrics-certs\") pod \"controller-68d546b9d8-hsbdv\" (UID: \"37d57598-c98a-45ff-8de0-c51a5e39adeb\") " pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.186454 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5638abcc-fdb8-4ed6-a630-8562d79668a5-frr-sockets\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.187009 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5638abcc-fdb8-4ed6-a630-8562d79668a5-frr-sockets\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: E1002 09:54:39.187181 4771 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 02 09:54:39 crc kubenswrapper[4771]: E1002 09:54:39.187237 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3c832819-5b80-4548-8941-6e411bd1ba2f-cert podName:3c832819-5b80-4548-8941-6e411bd1ba2f nodeName:}" failed. No retries permitted until 2025-10-02 09:54:39.687215284 +0000 UTC m=+1067.334900361 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3c832819-5b80-4548-8941-6e411bd1ba2f-cert") pod "frr-k8s-webhook-server-64bf5d555-qvlg4" (UID: "3c832819-5b80-4548-8941-6e411bd1ba2f") : secret "frr-k8s-webhook-server-cert" not found Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.188798 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5638abcc-fdb8-4ed6-a630-8562d79668a5-frr-startup\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.189063 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5638abcc-fdb8-4ed6-a630-8562d79668a5-metrics\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.189290 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5638abcc-fdb8-4ed6-a630-8562d79668a5-reloader\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.197576 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5638abcc-fdb8-4ed6-a630-8562d79668a5-frr-conf\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.198919 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5638abcc-fdb8-4ed6-a630-8562d79668a5-metrics-certs\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.250631 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8bpv\" (UniqueName: \"kubernetes.io/projected/3c832819-5b80-4548-8941-6e411bd1ba2f-kube-api-access-c8bpv\") pod \"frr-k8s-webhook-server-64bf5d555-qvlg4\" (UID: \"3c832819-5b80-4548-8941-6e411bd1ba2f\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.272394 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpnh2\" (UniqueName: \"kubernetes.io/projected/5638abcc-fdb8-4ed6-a630-8562d79668a5-kube-api-access-fpnh2\") pod \"frr-k8s-cf6h7\" (UID: \"5638abcc-fdb8-4ed6-a630-8562d79668a5\") " pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.313039 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/381993c1-33fd-485c-bb38-3f18cfa91abf-memberlist\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.313110 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpj2j\" (UniqueName: \"kubernetes.io/projected/37d57598-c98a-45ff-8de0-c51a5e39adeb-kube-api-access-tpj2j\") pod \"controller-68d546b9d8-hsbdv\" (UID: \"37d57598-c98a-45ff-8de0-c51a5e39adeb\") " pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.313154 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/37d57598-c98a-45ff-8de0-c51a5e39adeb-cert\") pod \"controller-68d546b9d8-hsbdv\" (UID: \"37d57598-c98a-45ff-8de0-c51a5e39adeb\") " pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.313195 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/381993c1-33fd-485c-bb38-3f18cfa91abf-metrics-certs\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.313241 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/381993c1-33fd-485c-bb38-3f18cfa91abf-metallb-excludel2\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.313272 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37d57598-c98a-45ff-8de0-c51a5e39adeb-metrics-certs\") pod \"controller-68d546b9d8-hsbdv\" (UID: \"37d57598-c98a-45ff-8de0-c51a5e39adeb\") " pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.313318 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pl5d\" (UniqueName: \"kubernetes.io/projected/381993c1-33fd-485c-bb38-3f18cfa91abf-kube-api-access-7pl5d\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:39 crc kubenswrapper[4771]: E1002 09:54:39.313782 4771 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 09:54:39 crc kubenswrapper[4771]: E1002 09:54:39.313834 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/381993c1-33fd-485c-bb38-3f18cfa91abf-memberlist podName:381993c1-33fd-485c-bb38-3f18cfa91abf nodeName:}" failed. No retries permitted until 2025-10-02 09:54:39.813818664 +0000 UTC m=+1067.461503731 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/381993c1-33fd-485c-bb38-3f18cfa91abf-memberlist") pod "speaker-nrj2m" (UID: "381993c1-33fd-485c-bb38-3f18cfa91abf") : secret "metallb-memberlist" not found Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.315711 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/381993c1-33fd-485c-bb38-3f18cfa91abf-metallb-excludel2\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:39 crc kubenswrapper[4771]: E1002 09:54:39.316302 4771 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 02 09:54:39 crc kubenswrapper[4771]: E1002 09:54:39.316385 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/37d57598-c98a-45ff-8de0-c51a5e39adeb-metrics-certs podName:37d57598-c98a-45ff-8de0-c51a5e39adeb nodeName:}" failed. No retries permitted until 2025-10-02 09:54:39.816366826 +0000 UTC m=+1067.464051893 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/37d57598-c98a-45ff-8de0-c51a5e39adeb-metrics-certs") pod "controller-68d546b9d8-hsbdv" (UID: "37d57598-c98a-45ff-8de0-c51a5e39adeb") : secret "controller-certs-secret" not found Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.323753 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/37d57598-c98a-45ff-8de0-c51a5e39adeb-cert\") pod \"controller-68d546b9d8-hsbdv\" (UID: \"37d57598-c98a-45ff-8de0-c51a5e39adeb\") " pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.329542 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.352118 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpj2j\" (UniqueName: \"kubernetes.io/projected/37d57598-c98a-45ff-8de0-c51a5e39adeb-kube-api-access-tpj2j\") pod \"controller-68d546b9d8-hsbdv\" (UID: \"37d57598-c98a-45ff-8de0-c51a5e39adeb\") " pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.359014 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/381993c1-33fd-485c-bb38-3f18cfa91abf-metrics-certs\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.371154 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pl5d\" (UniqueName: \"kubernetes.io/projected/381993c1-33fd-485c-bb38-3f18cfa91abf-kube-api-access-7pl5d\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.637826 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cf6h7" event={"ID":"5638abcc-fdb8-4ed6-a630-8562d79668a5","Type":"ContainerStarted","Data":"71993eff4d9a78ebf8bb38723213b08ca4eb3b9d94be797e3042faf31afa50f3"} Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.721360 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3c832819-5b80-4548-8941-6e411bd1ba2f-cert\") pod \"frr-k8s-webhook-server-64bf5d555-qvlg4\" (UID: \"3c832819-5b80-4548-8941-6e411bd1ba2f\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.726219 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3c832819-5b80-4548-8941-6e411bd1ba2f-cert\") pod \"frr-k8s-webhook-server-64bf5d555-qvlg4\" (UID: \"3c832819-5b80-4548-8941-6e411bd1ba2f\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.823285 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/381993c1-33fd-485c-bb38-3f18cfa91abf-memberlist\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:39 crc kubenswrapper[4771]: E1002 09:54:39.823470 4771 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.823495 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37d57598-c98a-45ff-8de0-c51a5e39adeb-metrics-certs\") pod \"controller-68d546b9d8-hsbdv\" (UID: \"37d57598-c98a-45ff-8de0-c51a5e39adeb\") " pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:39 crc kubenswrapper[4771]: E1002 09:54:39.823546 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/381993c1-33fd-485c-bb38-3f18cfa91abf-memberlist podName:381993c1-33fd-485c-bb38-3f18cfa91abf nodeName:}" failed. No retries permitted until 2025-10-02 09:54:40.823523484 +0000 UTC m=+1068.471208571 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/381993c1-33fd-485c-bb38-3f18cfa91abf-memberlist") pod "speaker-nrj2m" (UID: "381993c1-33fd-485c-bb38-3f18cfa91abf") : secret "metallb-memberlist" not found Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.827765 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37d57598-c98a-45ff-8de0-c51a5e39adeb-metrics-certs\") pod \"controller-68d546b9d8-hsbdv\" (UID: \"37d57598-c98a-45ff-8de0-c51a5e39adeb\") " pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:39 crc kubenswrapper[4771]: I1002 09:54:39.944974 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" Oct 02 09:54:40 crc kubenswrapper[4771]: I1002 09:54:40.074472 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:40 crc kubenswrapper[4771]: I1002 09:54:40.376575 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4"] Oct 02 09:54:40 crc kubenswrapper[4771]: I1002 09:54:40.563773 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-hsbdv"] Oct 02 09:54:40 crc kubenswrapper[4771]: W1002 09:54:40.566434 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37d57598_c98a_45ff_8de0_c51a5e39adeb.slice/crio-f5a7d3e1550ae0e3428e5e6315e88342ef7bb8971f9f645fc18f8e6a1a0cce86 WatchSource:0}: Error finding container f5a7d3e1550ae0e3428e5e6315e88342ef7bb8971f9f645fc18f8e6a1a0cce86: Status 404 returned error can't find the container with id f5a7d3e1550ae0e3428e5e6315e88342ef7bb8971f9f645fc18f8e6a1a0cce86 Oct 02 09:54:40 crc kubenswrapper[4771]: I1002 09:54:40.644222 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" event={"ID":"3c832819-5b80-4548-8941-6e411bd1ba2f","Type":"ContainerStarted","Data":"02852db814143771b4f53e8bf22cd5254025cf0242a9908948764fdd2afc1c19"} Oct 02 09:54:40 crc kubenswrapper[4771]: I1002 09:54:40.645251 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-hsbdv" event={"ID":"37d57598-c98a-45ff-8de0-c51a5e39adeb","Type":"ContainerStarted","Data":"f5a7d3e1550ae0e3428e5e6315e88342ef7bb8971f9f645fc18f8e6a1a0cce86"} Oct 02 09:54:40 crc kubenswrapper[4771]: I1002 09:54:40.839597 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/381993c1-33fd-485c-bb38-3f18cfa91abf-memberlist\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:40 crc kubenswrapper[4771]: I1002 09:54:40.849669 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/381993c1-33fd-485c-bb38-3f18cfa91abf-memberlist\") pod \"speaker-nrj2m\" (UID: \"381993c1-33fd-485c-bb38-3f18cfa91abf\") " pod="metallb-system/speaker-nrj2m" Oct 02 09:54:40 crc kubenswrapper[4771]: I1002 09:54:40.953626 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-nrj2m" Oct 02 09:54:40 crc kubenswrapper[4771]: W1002 09:54:40.999244 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod381993c1_33fd_485c_bb38_3f18cfa91abf.slice/crio-bb041d8e73eb8a0ee44b0c03f6cda1a833721a034432ddd03aa4e103eeb9fa4e WatchSource:0}: Error finding container bb041d8e73eb8a0ee44b0c03f6cda1a833721a034432ddd03aa4e103eeb9fa4e: Status 404 returned error can't find the container with id bb041d8e73eb8a0ee44b0c03f6cda1a833721a034432ddd03aa4e103eeb9fa4e Oct 02 09:54:41 crc kubenswrapper[4771]: I1002 09:54:41.657818 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nrj2m" event={"ID":"381993c1-33fd-485c-bb38-3f18cfa91abf","Type":"ContainerStarted","Data":"2b42ac8982a9349044c4186fe637fa6dfd4a41584a5f147af72dcf4105ceed6d"} Oct 02 09:54:41 crc kubenswrapper[4771]: I1002 09:54:41.658172 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nrj2m" event={"ID":"381993c1-33fd-485c-bb38-3f18cfa91abf","Type":"ContainerStarted","Data":"bb041d8e73eb8a0ee44b0c03f6cda1a833721a034432ddd03aa4e103eeb9fa4e"} Oct 02 09:54:41 crc kubenswrapper[4771]: I1002 09:54:41.663097 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-hsbdv" event={"ID":"37d57598-c98a-45ff-8de0-c51a5e39adeb","Type":"ContainerStarted","Data":"a66e017e03a85dea3acafe855edacf3e7bb5932e5e940ffe8197d32ed17e2c58"} Oct 02 09:54:41 crc kubenswrapper[4771]: I1002 09:54:41.663177 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-hsbdv" event={"ID":"37d57598-c98a-45ff-8de0-c51a5e39adeb","Type":"ContainerStarted","Data":"9141253852b9cf0402bdf7be66c7c58630c82f42d196e81c0280326baaa2719c"} Oct 02 09:54:41 crc kubenswrapper[4771]: I1002 09:54:41.663298 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:41 crc kubenswrapper[4771]: I1002 09:54:41.703642 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-hsbdv" podStartSLOduration=2.703607851 podStartE2EDuration="2.703607851s" podCreationTimestamp="2025-10-02 09:54:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:54:41.685583711 +0000 UTC m=+1069.333268778" watchObservedRunningTime="2025-10-02 09:54:41.703607851 +0000 UTC m=+1069.351292918" Oct 02 09:54:42 crc kubenswrapper[4771]: I1002 09:54:42.677043 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nrj2m" event={"ID":"381993c1-33fd-485c-bb38-3f18cfa91abf","Type":"ContainerStarted","Data":"db2bc4987234cf40c9ebf87a3a7f0f8c29713fea564b1e83965b2f6174a4bec6"} Oct 02 09:54:42 crc kubenswrapper[4771]: I1002 09:54:42.677465 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-nrj2m" Oct 02 09:54:42 crc kubenswrapper[4771]: I1002 09:54:42.696362 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-nrj2m" podStartSLOduration=3.69633881 podStartE2EDuration="3.69633881s" podCreationTimestamp="2025-10-02 09:54:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:54:42.693384288 +0000 UTC m=+1070.341069375" watchObservedRunningTime="2025-10-02 09:54:42.69633881 +0000 UTC m=+1070.344023887" Oct 02 09:54:47 crc kubenswrapper[4771]: I1002 09:54:47.729415 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" event={"ID":"3c832819-5b80-4548-8941-6e411bd1ba2f","Type":"ContainerStarted","Data":"e3012acae64a26c98b31609e94458f2a25f1e093f47c27ac917c4a4e85107508"} Oct 02 09:54:47 crc kubenswrapper[4771]: I1002 09:54:47.730027 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" Oct 02 09:54:47 crc kubenswrapper[4771]: I1002 09:54:47.731623 4771 generic.go:334] "Generic (PLEG): container finished" podID="5638abcc-fdb8-4ed6-a630-8562d79668a5" containerID="29fac951623da328468fd8ef3badfd9ba4b98da7bff66ec9d77c7951b78ed366" exitCode=0 Oct 02 09:54:47 crc kubenswrapper[4771]: I1002 09:54:47.731656 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cf6h7" event={"ID":"5638abcc-fdb8-4ed6-a630-8562d79668a5","Type":"ContainerDied","Data":"29fac951623da328468fd8ef3badfd9ba4b98da7bff66ec9d77c7951b78ed366"} Oct 02 09:54:47 crc kubenswrapper[4771]: I1002 09:54:47.758518 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" podStartSLOduration=2.783403802 podStartE2EDuration="9.758502862s" podCreationTimestamp="2025-10-02 09:54:38 +0000 UTC" firstStartedPulling="2025-10-02 09:54:40.378977041 +0000 UTC m=+1068.026662108" lastFinishedPulling="2025-10-02 09:54:47.354076101 +0000 UTC m=+1075.001761168" observedRunningTime="2025-10-02 09:54:47.754208717 +0000 UTC m=+1075.401893784" watchObservedRunningTime="2025-10-02 09:54:47.758502862 +0000 UTC m=+1075.406187929" Oct 02 09:54:48 crc kubenswrapper[4771]: I1002 09:54:48.740800 4771 generic.go:334] "Generic (PLEG): container finished" podID="5638abcc-fdb8-4ed6-a630-8562d79668a5" containerID="a0b6b7827892065ee20791c216786abbc20e8e2ec1d09804cdb9961f65f66888" exitCode=0 Oct 02 09:54:48 crc kubenswrapper[4771]: I1002 09:54:48.741230 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cf6h7" event={"ID":"5638abcc-fdb8-4ed6-a630-8562d79668a5","Type":"ContainerDied","Data":"a0b6b7827892065ee20791c216786abbc20e8e2ec1d09804cdb9961f65f66888"} Oct 02 09:54:49 crc kubenswrapper[4771]: I1002 09:54:49.751924 4771 generic.go:334] "Generic (PLEG): container finished" podID="5638abcc-fdb8-4ed6-a630-8562d79668a5" containerID="0b8e04c234790a6b85b185c6c1ec5d542386861faec85335f14c6754d29d8e30" exitCode=0 Oct 02 09:54:49 crc kubenswrapper[4771]: I1002 09:54:49.751976 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cf6h7" event={"ID":"5638abcc-fdb8-4ed6-a630-8562d79668a5","Type":"ContainerDied","Data":"0b8e04c234790a6b85b185c6c1ec5d542386861faec85335f14c6754d29d8e30"} Oct 02 09:54:50 crc kubenswrapper[4771]: I1002 09:54:50.080697 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-hsbdv" Oct 02 09:54:50 crc kubenswrapper[4771]: I1002 09:54:50.766208 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cf6h7" event={"ID":"5638abcc-fdb8-4ed6-a630-8562d79668a5","Type":"ContainerStarted","Data":"4435788251f219725268ee90dd2e4cef3d52f861d5101a50f45d30758c855218"} Oct 02 09:54:50 crc kubenswrapper[4771]: I1002 09:54:50.766859 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cf6h7" event={"ID":"5638abcc-fdb8-4ed6-a630-8562d79668a5","Type":"ContainerStarted","Data":"d08154cf7cfdea950ef6de4016f8af8d17e0ccb3947c0ab2b5c31c4c8da667b0"} Oct 02 09:54:50 crc kubenswrapper[4771]: I1002 09:54:50.766879 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cf6h7" event={"ID":"5638abcc-fdb8-4ed6-a630-8562d79668a5","Type":"ContainerStarted","Data":"6430f9bf2741beabdce433fa62dd622c28a53121dbd6a0c52485140335e7b90a"} Oct 02 09:54:50 crc kubenswrapper[4771]: I1002 09:54:50.766892 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cf6h7" event={"ID":"5638abcc-fdb8-4ed6-a630-8562d79668a5","Type":"ContainerStarted","Data":"cdc03532832055c1f779ecf74c51abf88cb2b41eda00acff897a7d5fcf80df11"} Oct 02 09:54:51 crc kubenswrapper[4771]: I1002 09:54:51.778740 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cf6h7" event={"ID":"5638abcc-fdb8-4ed6-a630-8562d79668a5","Type":"ContainerStarted","Data":"67dd32d71163eed87c13181a6273e909e6083243a7bb0f42498127fd9f414b4f"} Oct 02 09:54:51 crc kubenswrapper[4771]: I1002 09:54:51.779014 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:51 crc kubenswrapper[4771]: I1002 09:54:51.779025 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cf6h7" event={"ID":"5638abcc-fdb8-4ed6-a630-8562d79668a5","Type":"ContainerStarted","Data":"f3254b139ad80625dbb9680737b89145e72e267043f0a1ea73ce72c1697727b6"} Oct 02 09:54:51 crc kubenswrapper[4771]: I1002 09:54:51.800432 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-cf6h7" podStartSLOduration=5.946686597 podStartE2EDuration="13.800415751s" podCreationTimestamp="2025-10-02 09:54:38 +0000 UTC" firstStartedPulling="2025-10-02 09:54:39.491384718 +0000 UTC m=+1067.139069775" lastFinishedPulling="2025-10-02 09:54:47.345113872 +0000 UTC m=+1074.992798929" observedRunningTime="2025-10-02 09:54:51.799537669 +0000 UTC m=+1079.447222736" watchObservedRunningTime="2025-10-02 09:54:51.800415751 +0000 UTC m=+1079.448100818" Oct 02 09:54:54 crc kubenswrapper[4771]: I1002 09:54:54.330421 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:54 crc kubenswrapper[4771]: I1002 09:54:54.379707 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:59 crc kubenswrapper[4771]: I1002 09:54:59.335083 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-cf6h7" Oct 02 09:54:59 crc kubenswrapper[4771]: I1002 09:54:59.952618 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qvlg4" Oct 02 09:55:00 crc kubenswrapper[4771]: I1002 09:55:00.957832 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-nrj2m" Oct 02 09:55:04 crc kubenswrapper[4771]: I1002 09:55:04.066247 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2gbmk"] Oct 02 09:55:04 crc kubenswrapper[4771]: I1002 09:55:04.067857 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2gbmk" Oct 02 09:55:04 crc kubenswrapper[4771]: I1002 09:55:04.069972 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 02 09:55:04 crc kubenswrapper[4771]: I1002 09:55:04.070083 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 02 09:55:04 crc kubenswrapper[4771]: I1002 09:55:04.072071 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-gzdcp" Oct 02 09:55:04 crc kubenswrapper[4771]: I1002 09:55:04.101639 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2gbmk"] Oct 02 09:55:04 crc kubenswrapper[4771]: I1002 09:55:04.202820 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcdcn\" (UniqueName: \"kubernetes.io/projected/2756f3a0-27ac-43c1-8371-423813fc5523-kube-api-access-dcdcn\") pod \"openstack-operator-index-2gbmk\" (UID: \"2756f3a0-27ac-43c1-8371-423813fc5523\") " pod="openstack-operators/openstack-operator-index-2gbmk" Oct 02 09:55:04 crc kubenswrapper[4771]: I1002 09:55:04.304581 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcdcn\" (UniqueName: \"kubernetes.io/projected/2756f3a0-27ac-43c1-8371-423813fc5523-kube-api-access-dcdcn\") pod \"openstack-operator-index-2gbmk\" (UID: \"2756f3a0-27ac-43c1-8371-423813fc5523\") " pod="openstack-operators/openstack-operator-index-2gbmk" Oct 02 09:55:04 crc kubenswrapper[4771]: I1002 09:55:04.328067 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcdcn\" (UniqueName: \"kubernetes.io/projected/2756f3a0-27ac-43c1-8371-423813fc5523-kube-api-access-dcdcn\") pod \"openstack-operator-index-2gbmk\" (UID: \"2756f3a0-27ac-43c1-8371-423813fc5523\") " pod="openstack-operators/openstack-operator-index-2gbmk" Oct 02 09:55:04 crc kubenswrapper[4771]: I1002 09:55:04.388350 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2gbmk" Oct 02 09:55:04 crc kubenswrapper[4771]: I1002 09:55:04.853342 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2gbmk"] Oct 02 09:55:04 crc kubenswrapper[4771]: I1002 09:55:04.888555 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2gbmk" event={"ID":"2756f3a0-27ac-43c1-8371-423813fc5523","Type":"ContainerStarted","Data":"98db63d3b3bcf3ad776f26b96aec719f5e67c329f5e766096e4e94a12b56793a"} Oct 02 09:55:07 crc kubenswrapper[4771]: I1002 09:55:07.445487 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2gbmk"] Oct 02 09:55:07 crc kubenswrapper[4771]: I1002 09:55:07.919984 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2gbmk" event={"ID":"2756f3a0-27ac-43c1-8371-423813fc5523","Type":"ContainerStarted","Data":"ebc5beb8d1a2766998cf0796dab4590af9e9ef7158f10bbf0abe03f59801df30"} Oct 02 09:55:07 crc kubenswrapper[4771]: I1002 09:55:07.940536 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2gbmk" podStartSLOduration=1.872367881 podStartE2EDuration="3.940518247s" podCreationTimestamp="2025-10-02 09:55:04 +0000 UTC" firstStartedPulling="2025-10-02 09:55:04.868271264 +0000 UTC m=+1092.515956331" lastFinishedPulling="2025-10-02 09:55:06.93642163 +0000 UTC m=+1094.584106697" observedRunningTime="2025-10-02 09:55:07.936603011 +0000 UTC m=+1095.584288078" watchObservedRunningTime="2025-10-02 09:55:07.940518247 +0000 UTC m=+1095.588203314" Oct 02 09:55:08 crc kubenswrapper[4771]: I1002 09:55:08.047065 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-cpqp2"] Oct 02 09:55:08 crc kubenswrapper[4771]: I1002 09:55:08.048338 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-cpqp2" Oct 02 09:55:08 crc kubenswrapper[4771]: I1002 09:55:08.057722 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-cpqp2"] Oct 02 09:55:08 crc kubenswrapper[4771]: I1002 09:55:08.176378 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg9r8\" (UniqueName: \"kubernetes.io/projected/46c1bd11-aba5-4bc3-ae7e-fa24dd6687f6-kube-api-access-dg9r8\") pod \"openstack-operator-index-cpqp2\" (UID: \"46c1bd11-aba5-4bc3-ae7e-fa24dd6687f6\") " pod="openstack-operators/openstack-operator-index-cpqp2" Oct 02 09:55:08 crc kubenswrapper[4771]: I1002 09:55:08.278485 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg9r8\" (UniqueName: \"kubernetes.io/projected/46c1bd11-aba5-4bc3-ae7e-fa24dd6687f6-kube-api-access-dg9r8\") pod \"openstack-operator-index-cpqp2\" (UID: \"46c1bd11-aba5-4bc3-ae7e-fa24dd6687f6\") " pod="openstack-operators/openstack-operator-index-cpqp2" Oct 02 09:55:08 crc kubenswrapper[4771]: I1002 09:55:08.299220 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg9r8\" (UniqueName: \"kubernetes.io/projected/46c1bd11-aba5-4bc3-ae7e-fa24dd6687f6-kube-api-access-dg9r8\") pod \"openstack-operator-index-cpqp2\" (UID: \"46c1bd11-aba5-4bc3-ae7e-fa24dd6687f6\") " pod="openstack-operators/openstack-operator-index-cpqp2" Oct 02 09:55:08 crc kubenswrapper[4771]: I1002 09:55:08.375437 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-cpqp2" Oct 02 09:55:08 crc kubenswrapper[4771]: I1002 09:55:08.778019 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-cpqp2"] Oct 02 09:55:08 crc kubenswrapper[4771]: I1002 09:55:08.928617 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-cpqp2" event={"ID":"46c1bd11-aba5-4bc3-ae7e-fa24dd6687f6","Type":"ContainerStarted","Data":"97e6c4f5b4254186e94c57703318b1d75a2949fb8caeac7daf2fbdf65cce1d5b"} Oct 02 09:55:08 crc kubenswrapper[4771]: I1002 09:55:08.928761 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-2gbmk" podUID="2756f3a0-27ac-43c1-8371-423813fc5523" containerName="registry-server" containerID="cri-o://ebc5beb8d1a2766998cf0796dab4590af9e9ef7158f10bbf0abe03f59801df30" gracePeriod=2 Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.440616 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2gbmk" Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.602481 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcdcn\" (UniqueName: \"kubernetes.io/projected/2756f3a0-27ac-43c1-8371-423813fc5523-kube-api-access-dcdcn\") pod \"2756f3a0-27ac-43c1-8371-423813fc5523\" (UID: \"2756f3a0-27ac-43c1-8371-423813fc5523\") " Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.609026 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2756f3a0-27ac-43c1-8371-423813fc5523-kube-api-access-dcdcn" (OuterVolumeSpecName: "kube-api-access-dcdcn") pod "2756f3a0-27ac-43c1-8371-423813fc5523" (UID: "2756f3a0-27ac-43c1-8371-423813fc5523"). InnerVolumeSpecName "kube-api-access-dcdcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.705020 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcdcn\" (UniqueName: \"kubernetes.io/projected/2756f3a0-27ac-43c1-8371-423813fc5523-kube-api-access-dcdcn\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.936593 4771 generic.go:334] "Generic (PLEG): container finished" podID="2756f3a0-27ac-43c1-8371-423813fc5523" containerID="ebc5beb8d1a2766998cf0796dab4590af9e9ef7158f10bbf0abe03f59801df30" exitCode=0 Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.936638 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2gbmk" Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.936671 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2gbmk" event={"ID":"2756f3a0-27ac-43c1-8371-423813fc5523","Type":"ContainerDied","Data":"ebc5beb8d1a2766998cf0796dab4590af9e9ef7158f10bbf0abe03f59801df30"} Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.936865 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2gbmk" event={"ID":"2756f3a0-27ac-43c1-8371-423813fc5523","Type":"ContainerDied","Data":"98db63d3b3bcf3ad776f26b96aec719f5e67c329f5e766096e4e94a12b56793a"} Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.936914 4771 scope.go:117] "RemoveContainer" containerID="ebc5beb8d1a2766998cf0796dab4590af9e9ef7158f10bbf0abe03f59801df30" Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.938212 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-cpqp2" event={"ID":"46c1bd11-aba5-4bc3-ae7e-fa24dd6687f6","Type":"ContainerStarted","Data":"7e2536e507536ba16c75c6379a4752440289fe02f1087b529c950dee88fc6c6a"} Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.958266 4771 scope.go:117] "RemoveContainer" containerID="ebc5beb8d1a2766998cf0796dab4590af9e9ef7158f10bbf0abe03f59801df30" Oct 02 09:55:09 crc kubenswrapper[4771]: E1002 09:55:09.958664 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebc5beb8d1a2766998cf0796dab4590af9e9ef7158f10bbf0abe03f59801df30\": container with ID starting with ebc5beb8d1a2766998cf0796dab4590af9e9ef7158f10bbf0abe03f59801df30 not found: ID does not exist" containerID="ebc5beb8d1a2766998cf0796dab4590af9e9ef7158f10bbf0abe03f59801df30" Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.958703 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebc5beb8d1a2766998cf0796dab4590af9e9ef7158f10bbf0abe03f59801df30"} err="failed to get container status \"ebc5beb8d1a2766998cf0796dab4590af9e9ef7158f10bbf0abe03f59801df30\": rpc error: code = NotFound desc = could not find container \"ebc5beb8d1a2766998cf0796dab4590af9e9ef7158f10bbf0abe03f59801df30\": container with ID starting with ebc5beb8d1a2766998cf0796dab4590af9e9ef7158f10bbf0abe03f59801df30 not found: ID does not exist" Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.958862 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-cpqp2" podStartSLOduration=1.9034650050000002 podStartE2EDuration="1.958818476s" podCreationTimestamp="2025-10-02 09:55:08 +0000 UTC" firstStartedPulling="2025-10-02 09:55:08.782873396 +0000 UTC m=+1096.430558453" lastFinishedPulling="2025-10-02 09:55:08.838226857 +0000 UTC m=+1096.485911924" observedRunningTime="2025-10-02 09:55:09.956824688 +0000 UTC m=+1097.604509755" watchObservedRunningTime="2025-10-02 09:55:09.958818476 +0000 UTC m=+1097.606503543" Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.980080 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2gbmk"] Oct 02 09:55:09 crc kubenswrapper[4771]: I1002 09:55:09.988630 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-2gbmk"] Oct 02 09:55:11 crc kubenswrapper[4771]: I1002 09:55:11.691200 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2756f3a0-27ac-43c1-8371-423813fc5523" path="/var/lib/kubelet/pods/2756f3a0-27ac-43c1-8371-423813fc5523/volumes" Oct 02 09:55:12 crc kubenswrapper[4771]: I1002 09:55:12.146758 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:55:12 crc kubenswrapper[4771]: I1002 09:55:12.147093 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:55:18 crc kubenswrapper[4771]: I1002 09:55:18.376205 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-cpqp2" Oct 02 09:55:18 crc kubenswrapper[4771]: I1002 09:55:18.377234 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-cpqp2" Oct 02 09:55:18 crc kubenswrapper[4771]: I1002 09:55:18.409911 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-cpqp2" Oct 02 09:55:19 crc kubenswrapper[4771]: I1002 09:55:19.052258 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-cpqp2" Oct 02 09:55:19 crc kubenswrapper[4771]: I1002 09:55:19.934619 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq"] Oct 02 09:55:19 crc kubenswrapper[4771]: E1002 09:55:19.935519 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2756f3a0-27ac-43c1-8371-423813fc5523" containerName="registry-server" Oct 02 09:55:19 crc kubenswrapper[4771]: I1002 09:55:19.935534 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2756f3a0-27ac-43c1-8371-423813fc5523" containerName="registry-server" Oct 02 09:55:19 crc kubenswrapper[4771]: I1002 09:55:19.935737 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2756f3a0-27ac-43c1-8371-423813fc5523" containerName="registry-server" Oct 02 09:55:19 crc kubenswrapper[4771]: I1002 09:55:19.937016 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" Oct 02 09:55:19 crc kubenswrapper[4771]: I1002 09:55:19.939792 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-qc459" Oct 02 09:55:19 crc kubenswrapper[4771]: I1002 09:55:19.944848 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq"] Oct 02 09:55:20 crc kubenswrapper[4771]: I1002 09:55:20.028467 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngfwv\" (UniqueName: \"kubernetes.io/projected/08da521d-cddc-4af0-922b-673e392a68f3-kube-api-access-ngfwv\") pod \"838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq\" (UID: \"08da521d-cddc-4af0-922b-673e392a68f3\") " pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" Oct 02 09:55:20 crc kubenswrapper[4771]: I1002 09:55:20.028565 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08da521d-cddc-4af0-922b-673e392a68f3-util\") pod \"838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq\" (UID: \"08da521d-cddc-4af0-922b-673e392a68f3\") " pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" Oct 02 09:55:20 crc kubenswrapper[4771]: I1002 09:55:20.028603 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08da521d-cddc-4af0-922b-673e392a68f3-bundle\") pod \"838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq\" (UID: \"08da521d-cddc-4af0-922b-673e392a68f3\") " pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" Oct 02 09:55:20 crc kubenswrapper[4771]: I1002 09:55:20.130583 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngfwv\" (UniqueName: \"kubernetes.io/projected/08da521d-cddc-4af0-922b-673e392a68f3-kube-api-access-ngfwv\") pod \"838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq\" (UID: \"08da521d-cddc-4af0-922b-673e392a68f3\") " pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" Oct 02 09:55:20 crc kubenswrapper[4771]: I1002 09:55:20.130662 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08da521d-cddc-4af0-922b-673e392a68f3-util\") pod \"838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq\" (UID: \"08da521d-cddc-4af0-922b-673e392a68f3\") " pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" Oct 02 09:55:20 crc kubenswrapper[4771]: I1002 09:55:20.130689 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08da521d-cddc-4af0-922b-673e392a68f3-bundle\") pod \"838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq\" (UID: \"08da521d-cddc-4af0-922b-673e392a68f3\") " pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" Oct 02 09:55:20 crc kubenswrapper[4771]: I1002 09:55:20.131245 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08da521d-cddc-4af0-922b-673e392a68f3-util\") pod \"838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq\" (UID: \"08da521d-cddc-4af0-922b-673e392a68f3\") " pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" Oct 02 09:55:20 crc kubenswrapper[4771]: I1002 09:55:20.131304 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08da521d-cddc-4af0-922b-673e392a68f3-bundle\") pod \"838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq\" (UID: \"08da521d-cddc-4af0-922b-673e392a68f3\") " pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" Oct 02 09:55:20 crc kubenswrapper[4771]: I1002 09:55:20.151013 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngfwv\" (UniqueName: \"kubernetes.io/projected/08da521d-cddc-4af0-922b-673e392a68f3-kube-api-access-ngfwv\") pod \"838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq\" (UID: \"08da521d-cddc-4af0-922b-673e392a68f3\") " pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" Oct 02 09:55:20 crc kubenswrapper[4771]: I1002 09:55:20.257466 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" Oct 02 09:55:20 crc kubenswrapper[4771]: I1002 09:55:20.667700 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq"] Oct 02 09:55:21 crc kubenswrapper[4771]: I1002 09:55:21.039567 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" event={"ID":"08da521d-cddc-4af0-922b-673e392a68f3","Type":"ContainerStarted","Data":"23fd18eb7561ef232de5db1198d888d901e64320d254cdc85cc8964b07808b6a"} Oct 02 09:55:22 crc kubenswrapper[4771]: I1002 09:55:22.048786 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" event={"ID":"08da521d-cddc-4af0-922b-673e392a68f3","Type":"ContainerStarted","Data":"ccec2094d80b1e3a5dd46ce93085bc1447ec12481c01165794df3bc9d16d486e"} Oct 02 09:55:23 crc kubenswrapper[4771]: I1002 09:55:23.057747 4771 generic.go:334] "Generic (PLEG): container finished" podID="08da521d-cddc-4af0-922b-673e392a68f3" containerID="ccec2094d80b1e3a5dd46ce93085bc1447ec12481c01165794df3bc9d16d486e" exitCode=0 Oct 02 09:55:23 crc kubenswrapper[4771]: I1002 09:55:23.057799 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" event={"ID":"08da521d-cddc-4af0-922b-673e392a68f3","Type":"ContainerDied","Data":"ccec2094d80b1e3a5dd46ce93085bc1447ec12481c01165794df3bc9d16d486e"} Oct 02 09:55:25 crc kubenswrapper[4771]: I1002 09:55:25.072816 4771 generic.go:334] "Generic (PLEG): container finished" podID="08da521d-cddc-4af0-922b-673e392a68f3" containerID="0271f89490481eb91883affab4bd3cfd3ee152ded381122143ca663ac7b17fe7" exitCode=0 Oct 02 09:55:25 crc kubenswrapper[4771]: I1002 09:55:25.072888 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" event={"ID":"08da521d-cddc-4af0-922b-673e392a68f3","Type":"ContainerDied","Data":"0271f89490481eb91883affab4bd3cfd3ee152ded381122143ca663ac7b17fe7"} Oct 02 09:55:26 crc kubenswrapper[4771]: I1002 09:55:26.084331 4771 generic.go:334] "Generic (PLEG): container finished" podID="08da521d-cddc-4af0-922b-673e392a68f3" containerID="ae969f93a6f3001387d4b6f92a08a5ecd4fe7947330b31d52860ea0fdef0ca3b" exitCode=0 Oct 02 09:55:26 crc kubenswrapper[4771]: I1002 09:55:26.084437 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" event={"ID":"08da521d-cddc-4af0-922b-673e392a68f3","Type":"ContainerDied","Data":"ae969f93a6f3001387d4b6f92a08a5ecd4fe7947330b31d52860ea0fdef0ca3b"} Oct 02 09:55:27 crc kubenswrapper[4771]: I1002 09:55:27.406907 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" Oct 02 09:55:27 crc kubenswrapper[4771]: I1002 09:55:27.557780 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngfwv\" (UniqueName: \"kubernetes.io/projected/08da521d-cddc-4af0-922b-673e392a68f3-kube-api-access-ngfwv\") pod \"08da521d-cddc-4af0-922b-673e392a68f3\" (UID: \"08da521d-cddc-4af0-922b-673e392a68f3\") " Oct 02 09:55:27 crc kubenswrapper[4771]: I1002 09:55:27.557852 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08da521d-cddc-4af0-922b-673e392a68f3-util\") pod \"08da521d-cddc-4af0-922b-673e392a68f3\" (UID: \"08da521d-cddc-4af0-922b-673e392a68f3\") " Oct 02 09:55:27 crc kubenswrapper[4771]: I1002 09:55:27.557935 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08da521d-cddc-4af0-922b-673e392a68f3-bundle\") pod \"08da521d-cddc-4af0-922b-673e392a68f3\" (UID: \"08da521d-cddc-4af0-922b-673e392a68f3\") " Oct 02 09:55:27 crc kubenswrapper[4771]: I1002 09:55:27.558722 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08da521d-cddc-4af0-922b-673e392a68f3-bundle" (OuterVolumeSpecName: "bundle") pod "08da521d-cddc-4af0-922b-673e392a68f3" (UID: "08da521d-cddc-4af0-922b-673e392a68f3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:55:27 crc kubenswrapper[4771]: I1002 09:55:27.563959 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08da521d-cddc-4af0-922b-673e392a68f3-kube-api-access-ngfwv" (OuterVolumeSpecName: "kube-api-access-ngfwv") pod "08da521d-cddc-4af0-922b-673e392a68f3" (UID: "08da521d-cddc-4af0-922b-673e392a68f3"). InnerVolumeSpecName "kube-api-access-ngfwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:55:27 crc kubenswrapper[4771]: I1002 09:55:27.659768 4771 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08da521d-cddc-4af0-922b-673e392a68f3-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:27 crc kubenswrapper[4771]: I1002 09:55:27.659811 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngfwv\" (UniqueName: \"kubernetes.io/projected/08da521d-cddc-4af0-922b-673e392a68f3-kube-api-access-ngfwv\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:27 crc kubenswrapper[4771]: I1002 09:55:27.715780 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08da521d-cddc-4af0-922b-673e392a68f3-util" (OuterVolumeSpecName: "util") pod "08da521d-cddc-4af0-922b-673e392a68f3" (UID: "08da521d-cddc-4af0-922b-673e392a68f3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:55:27 crc kubenswrapper[4771]: I1002 09:55:27.761717 4771 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08da521d-cddc-4af0-922b-673e392a68f3-util\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:28 crc kubenswrapper[4771]: I1002 09:55:28.101786 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" event={"ID":"08da521d-cddc-4af0-922b-673e392a68f3","Type":"ContainerDied","Data":"23fd18eb7561ef232de5db1198d888d901e64320d254cdc85cc8964b07808b6a"} Oct 02 09:55:28 crc kubenswrapper[4771]: I1002 09:55:28.101830 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23fd18eb7561ef232de5db1198d888d901e64320d254cdc85cc8964b07808b6a" Oct 02 09:55:28 crc kubenswrapper[4771]: I1002 09:55:28.101870 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq" Oct 02 09:55:32 crc kubenswrapper[4771]: I1002 09:55:32.620116 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7"] Oct 02 09:55:32 crc kubenswrapper[4771]: E1002 09:55:32.621483 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08da521d-cddc-4af0-922b-673e392a68f3" containerName="pull" Oct 02 09:55:32 crc kubenswrapper[4771]: I1002 09:55:32.621507 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="08da521d-cddc-4af0-922b-673e392a68f3" containerName="pull" Oct 02 09:55:32 crc kubenswrapper[4771]: E1002 09:55:32.621574 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08da521d-cddc-4af0-922b-673e392a68f3" containerName="extract" Oct 02 09:55:32 crc kubenswrapper[4771]: I1002 09:55:32.621583 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="08da521d-cddc-4af0-922b-673e392a68f3" containerName="extract" Oct 02 09:55:32 crc kubenswrapper[4771]: E1002 09:55:32.621611 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08da521d-cddc-4af0-922b-673e392a68f3" containerName="util" Oct 02 09:55:32 crc kubenswrapper[4771]: I1002 09:55:32.621618 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="08da521d-cddc-4af0-922b-673e392a68f3" containerName="util" Oct 02 09:55:32 crc kubenswrapper[4771]: I1002 09:55:32.621813 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="08da521d-cddc-4af0-922b-673e392a68f3" containerName="extract" Oct 02 09:55:32 crc kubenswrapper[4771]: I1002 09:55:32.622922 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7" Oct 02 09:55:32 crc kubenswrapper[4771]: I1002 09:55:32.628794 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-xkl4f" Oct 02 09:55:32 crc kubenswrapper[4771]: I1002 09:55:32.645572 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7"] Oct 02 09:55:32 crc kubenswrapper[4771]: I1002 09:55:32.759669 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk8v5\" (UniqueName: \"kubernetes.io/projected/f01b0e6f-9cfa-45ac-878d-4af22bea031c-kube-api-access-wk8v5\") pod \"openstack-operator-controller-operator-744d8c869b-vknn7\" (UID: \"f01b0e6f-9cfa-45ac-878d-4af22bea031c\") " pod="openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7" Oct 02 09:55:32 crc kubenswrapper[4771]: I1002 09:55:32.862000 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk8v5\" (UniqueName: \"kubernetes.io/projected/f01b0e6f-9cfa-45ac-878d-4af22bea031c-kube-api-access-wk8v5\") pod \"openstack-operator-controller-operator-744d8c869b-vknn7\" (UID: \"f01b0e6f-9cfa-45ac-878d-4af22bea031c\") " pod="openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7" Oct 02 09:55:32 crc kubenswrapper[4771]: I1002 09:55:32.883591 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk8v5\" (UniqueName: \"kubernetes.io/projected/f01b0e6f-9cfa-45ac-878d-4af22bea031c-kube-api-access-wk8v5\") pod \"openstack-operator-controller-operator-744d8c869b-vknn7\" (UID: \"f01b0e6f-9cfa-45ac-878d-4af22bea031c\") " pod="openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7" Oct 02 09:55:32 crc kubenswrapper[4771]: I1002 09:55:32.950277 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7" Oct 02 09:55:33 crc kubenswrapper[4771]: I1002 09:55:33.438911 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7"] Oct 02 09:55:34 crc kubenswrapper[4771]: I1002 09:55:34.153789 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7" event={"ID":"f01b0e6f-9cfa-45ac-878d-4af22bea031c","Type":"ContainerStarted","Data":"c497b24812dc64fced67c2ed91e0c9205583000f1e04eba729f8d2e453861881"} Oct 02 09:55:39 crc kubenswrapper[4771]: I1002 09:55:39.204984 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7" event={"ID":"f01b0e6f-9cfa-45ac-878d-4af22bea031c","Type":"ContainerStarted","Data":"6a64202915f6567c955fdbd74f4c1aea229e570c22174df550d5f9186a1dbb6d"} Oct 02 09:55:42 crc kubenswrapper[4771]: I1002 09:55:42.146214 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:55:42 crc kubenswrapper[4771]: I1002 09:55:42.147199 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:55:42 crc kubenswrapper[4771]: I1002 09:55:42.239063 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7" event={"ID":"f01b0e6f-9cfa-45ac-878d-4af22bea031c","Type":"ContainerStarted","Data":"dd8a494148193640f8429f1a62d0afa78ad717dd3ecf35d2041a9bbbcb615def"} Oct 02 09:55:42 crc kubenswrapper[4771]: I1002 09:55:42.241329 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7" Oct 02 09:55:42 crc kubenswrapper[4771]: I1002 09:55:42.278933 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7" podStartSLOduration=2.247971912 podStartE2EDuration="10.278907939s" podCreationTimestamp="2025-10-02 09:55:32 +0000 UTC" firstStartedPulling="2025-10-02 09:55:33.447801781 +0000 UTC m=+1121.095486848" lastFinishedPulling="2025-10-02 09:55:41.478737808 +0000 UTC m=+1129.126422875" observedRunningTime="2025-10-02 09:55:42.271575151 +0000 UTC m=+1129.919260228" watchObservedRunningTime="2025-10-02 09:55:42.278907939 +0000 UTC m=+1129.926593006" Oct 02 09:55:43 crc kubenswrapper[4771]: I1002 09:55:43.248569 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-744d8c869b-vknn7" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.211963 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.213774 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.218629 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-5tf5d" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.231862 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.237764 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.240170 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.243936 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-5jcp6" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.248259 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.250009 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.262975 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-2tvxd" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.277371 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.286083 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.295822 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.297410 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.302020 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-5865t" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.347106 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rksw\" (UniqueName: \"kubernetes.io/projected/f0638dc6-d014-4a78-bcfa-eb4e5cea5caf-kube-api-access-2rksw\") pod \"designate-operator-controller-manager-84f4f7b77b-5xvqw\" (UID: \"f0638dc6-d014-4a78-bcfa-eb4e5cea5caf\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.348937 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2vwm\" (UniqueName: \"kubernetes.io/projected/5fcc6b1c-2163-454b-9172-16852b9f8966-kube-api-access-q2vwm\") pod \"cinder-operator-controller-manager-644bddb6d8-dmwmw\" (UID: \"5fcc6b1c-2163-454b-9172-16852b9f8966\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.349038 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pjpv\" (UniqueName: \"kubernetes.io/projected/4da5dda4-f0da-453d-ba80-a2fc95035688-kube-api-access-6pjpv\") pod \"barbican-operator-controller-manager-6ff8b75857-smc8h\" (UID: \"4da5dda4-f0da-453d-ba80-a2fc95035688\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.362695 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.366018 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.376417 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-97f6d" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.418261 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.450327 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhv6v\" (UniqueName: \"kubernetes.io/projected/6dc77080-a5ab-4db4-b150-2573589da332-kube-api-access-vhv6v\") pod \"glance-operator-controller-manager-84958c4d49-pnqnd\" (UID: \"6dc77080-a5ab-4db4-b150-2573589da332\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.450382 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pjpv\" (UniqueName: \"kubernetes.io/projected/4da5dda4-f0da-453d-ba80-a2fc95035688-kube-api-access-6pjpv\") pod \"barbican-operator-controller-manager-6ff8b75857-smc8h\" (UID: \"4da5dda4-f0da-453d-ba80-a2fc95035688\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.450467 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rksw\" (UniqueName: \"kubernetes.io/projected/f0638dc6-d014-4a78-bcfa-eb4e5cea5caf-kube-api-access-2rksw\") pod \"designate-operator-controller-manager-84f4f7b77b-5xvqw\" (UID: \"f0638dc6-d014-4a78-bcfa-eb4e5cea5caf\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.450517 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfzcj\" (UniqueName: \"kubernetes.io/projected/9804f0f6-0273-4b09-830f-f082cf718803-kube-api-access-vfzcj\") pod \"heat-operator-controller-manager-5d889d78cf-hh756\" (UID: \"9804f0f6-0273-4b09-830f-f082cf718803\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.450539 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2vwm\" (UniqueName: \"kubernetes.io/projected/5fcc6b1c-2163-454b-9172-16852b9f8966-kube-api-access-q2vwm\") pod \"cinder-operator-controller-manager-644bddb6d8-dmwmw\" (UID: \"5fcc6b1c-2163-454b-9172-16852b9f8966\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.455404 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.457367 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.502189 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-2scdd" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.504782 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.514235 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pjpv\" (UniqueName: \"kubernetes.io/projected/4da5dda4-f0da-453d-ba80-a2fc95035688-kube-api-access-6pjpv\") pod \"barbican-operator-controller-manager-6ff8b75857-smc8h\" (UID: \"4da5dda4-f0da-453d-ba80-a2fc95035688\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.518519 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.520667 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.523473 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-sm9db" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.524765 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.527403 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rksw\" (UniqueName: \"kubernetes.io/projected/f0638dc6-d014-4a78-bcfa-eb4e5cea5caf-kube-api-access-2rksw\") pod \"designate-operator-controller-manager-84f4f7b77b-5xvqw\" (UID: \"f0638dc6-d014-4a78-bcfa-eb4e5cea5caf\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.543673 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.547950 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2vwm\" (UniqueName: \"kubernetes.io/projected/5fcc6b1c-2163-454b-9172-16852b9f8966-kube-api-access-q2vwm\") pod \"cinder-operator-controller-manager-644bddb6d8-dmwmw\" (UID: \"5fcc6b1c-2163-454b-9172-16852b9f8966\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.564055 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfzcj\" (UniqueName: \"kubernetes.io/projected/9804f0f6-0273-4b09-830f-f082cf718803-kube-api-access-vfzcj\") pod \"heat-operator-controller-manager-5d889d78cf-hh756\" (UID: \"9804f0f6-0273-4b09-830f-f082cf718803\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.564151 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfgg4\" (UniqueName: \"kubernetes.io/projected/8e8d717b-41cc-4dd7-b9e3-4d06ac36cc54-kube-api-access-xfgg4\") pod \"horizon-operator-controller-manager-9f4696d94-lp8xj\" (UID: \"8e8d717b-41cc-4dd7-b9e3-4d06ac36cc54\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.564240 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhv6v\" (UniqueName: \"kubernetes.io/projected/6dc77080-a5ab-4db4-b150-2573589da332-kube-api-access-vhv6v\") pod \"glance-operator-controller-manager-84958c4d49-pnqnd\" (UID: \"6dc77080-a5ab-4db4-b150-2573589da332\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.565097 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.568098 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.585193 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.589620 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.611169 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhv6v\" (UniqueName: \"kubernetes.io/projected/6dc77080-a5ab-4db4-b150-2573589da332-kube-api-access-vhv6v\") pod \"glance-operator-controller-manager-84958c4d49-pnqnd\" (UID: \"6dc77080-a5ab-4db4-b150-2573589da332\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.620709 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.636402 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.652983 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-r24sp" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.653224 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.654001 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.655722 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.674605 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.677335 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.674710 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfgg4\" (UniqueName: \"kubernetes.io/projected/8e8d717b-41cc-4dd7-b9e3-4d06ac36cc54-kube-api-access-xfgg4\") pod \"horizon-operator-controller-manager-9f4696d94-lp8xj\" (UID: \"8e8d717b-41cc-4dd7-b9e3-4d06ac36cc54\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.678163 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5ds8\" (UniqueName: \"kubernetes.io/projected/e0df3275-79e0-49b8-93d5-585e2667a5c4-kube-api-access-d5ds8\") pod \"infra-operator-controller-manager-9d6c5db85-99mgr\" (UID: \"e0df3275-79e0-49b8-93d5-585e2667a5c4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.684119 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e0df3275-79e0-49b8-93d5-585e2667a5c4-cert\") pod \"infra-operator-controller-manager-9d6c5db85-99mgr\" (UID: \"e0df3275-79e0-49b8-93d5-585e2667a5c4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.672306 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfzcj\" (UniqueName: \"kubernetes.io/projected/9804f0f6-0273-4b09-830f-f082cf718803-kube-api-access-vfzcj\") pod \"heat-operator-controller-manager-5d889d78cf-hh756\" (UID: \"9804f0f6-0273-4b09-830f-f082cf718803\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.656201 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.678646 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.678513 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.694833 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.726981 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.727574 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-7htfl" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.727846 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-nd6rj" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.728012 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-5tp9z" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.736534 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.741217 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.742068 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.744903 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-xslwj" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.755875 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfgg4\" (UniqueName: \"kubernetes.io/projected/8e8d717b-41cc-4dd7-b9e3-4d06ac36cc54-kube-api-access-xfgg4\") pod \"horizon-operator-controller-manager-9f4696d94-lp8xj\" (UID: \"8e8d717b-41cc-4dd7-b9e3-4d06ac36cc54\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.764270 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.787049 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67zbb\" (UniqueName: \"kubernetes.io/projected/7c2eb266-7f51-45b9-8095-89d219bb59f0-kube-api-access-67zbb\") pod \"mariadb-operator-controller-manager-88c7-tgx7p\" (UID: \"7c2eb266-7f51-45b9-8095-89d219bb59f0\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.796560 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5njdl\" (UniqueName: \"kubernetes.io/projected/04a0b3f5-6e78-477a-8bd6-9250ba41eabf-kube-api-access-5njdl\") pod \"manila-operator-controller-manager-6d68dbc695-4dcd9\" (UID: \"04a0b3f5-6e78-477a-8bd6-9250ba41eabf\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.796664 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfph9\" (UniqueName: \"kubernetes.io/projected/7e625416-21a2-45e6-aec5-cf04c408e65e-kube-api-access-zfph9\") pod \"ironic-operator-controller-manager-5cd4858477-4nr9l\" (UID: \"7e625416-21a2-45e6-aec5-cf04c408e65e\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.796791 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5ds8\" (UniqueName: \"kubernetes.io/projected/e0df3275-79e0-49b8-93d5-585e2667a5c4-kube-api-access-d5ds8\") pod \"infra-operator-controller-manager-9d6c5db85-99mgr\" (UID: \"e0df3275-79e0-49b8-93d5-585e2667a5c4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.796949 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e0df3275-79e0-49b8-93d5-585e2667a5c4-cert\") pod \"infra-operator-controller-manager-9d6c5db85-99mgr\" (UID: \"e0df3275-79e0-49b8-93d5-585e2667a5c4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.797078 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvf25\" (UniqueName: \"kubernetes.io/projected/1f236f0a-744a-4f98-90e9-7c78ead31ddc-kube-api-access-wvf25\") pod \"keystone-operator-controller-manager-5bd55b4bff-jg6rv\" (UID: \"1f236f0a-744a-4f98-90e9-7c78ead31ddc\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" Oct 02 09:56:01 crc kubenswrapper[4771]: E1002 09:56:01.802512 4771 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 02 09:56:01 crc kubenswrapper[4771]: E1002 09:56:01.802581 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e0df3275-79e0-49b8-93d5-585e2667a5c4-cert podName:e0df3275-79e0-49b8-93d5-585e2667a5c4 nodeName:}" failed. No retries permitted until 2025-10-02 09:56:02.302555835 +0000 UTC m=+1149.950240902 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e0df3275-79e0-49b8-93d5-585e2667a5c4-cert") pod "infra-operator-controller-manager-9d6c5db85-99mgr" (UID: "e0df3275-79e0-49b8-93d5-585e2667a5c4") : secret "infra-operator-webhook-server-cert" not found Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.810254 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.820399 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.828606 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.829999 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.836344 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-jr229" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.845875 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.848197 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.860689 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-sq7d9" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.861806 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.862929 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5ds8\" (UniqueName: \"kubernetes.io/projected/e0df3275-79e0-49b8-93d5-585e2667a5c4-kube-api-access-d5ds8\") pod \"infra-operator-controller-manager-9d6c5db85-99mgr\" (UID: \"e0df3275-79e0-49b8-93d5-585e2667a5c4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.872757 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.885426 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.888877 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.891231 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.892745 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.892886 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-b2j96" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.893131 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.898494 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67zbb\" (UniqueName: \"kubernetes.io/projected/7c2eb266-7f51-45b9-8095-89d219bb59f0-kube-api-access-67zbb\") pod \"mariadb-operator-controller-manager-88c7-tgx7p\" (UID: \"7c2eb266-7f51-45b9-8095-89d219bb59f0\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.898635 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5njdl\" (UniqueName: \"kubernetes.io/projected/04a0b3f5-6e78-477a-8bd6-9250ba41eabf-kube-api-access-5njdl\") pod \"manila-operator-controller-manager-6d68dbc695-4dcd9\" (UID: \"04a0b3f5-6e78-477a-8bd6-9250ba41eabf\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.898678 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfph9\" (UniqueName: \"kubernetes.io/projected/7e625416-21a2-45e6-aec5-cf04c408e65e-kube-api-access-zfph9\") pod \"ironic-operator-controller-manager-5cd4858477-4nr9l\" (UID: \"7e625416-21a2-45e6-aec5-cf04c408e65e\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.898888 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lhp7\" (UniqueName: \"kubernetes.io/projected/4a399512-1d85-4f61-b8d4-800a57c33f4d-kube-api-access-7lhp7\") pod \"neutron-operator-controller-manager-849d5b9b84-zwjcd\" (UID: \"4a399512-1d85-4f61-b8d4-800a57c33f4d\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.898954 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvf25\" (UniqueName: \"kubernetes.io/projected/1f236f0a-744a-4f98-90e9-7c78ead31ddc-kube-api-access-wvf25\") pod \"keystone-operator-controller-manager-5bd55b4bff-jg6rv\" (UID: \"1f236f0a-744a-4f98-90e9-7c78ead31ddc\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.900973 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-qmt52" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.907199 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.909819 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d"] Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.952045 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfph9\" (UniqueName: \"kubernetes.io/projected/7e625416-21a2-45e6-aec5-cf04c408e65e-kube-api-access-zfph9\") pod \"ironic-operator-controller-manager-5cd4858477-4nr9l\" (UID: \"7e625416-21a2-45e6-aec5-cf04c408e65e\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.954989 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvf25\" (UniqueName: \"kubernetes.io/projected/1f236f0a-744a-4f98-90e9-7c78ead31ddc-kube-api-access-wvf25\") pod \"keystone-operator-controller-manager-5bd55b4bff-jg6rv\" (UID: \"1f236f0a-744a-4f98-90e9-7c78ead31ddc\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.959805 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67zbb\" (UniqueName: \"kubernetes.io/projected/7c2eb266-7f51-45b9-8095-89d219bb59f0-kube-api-access-67zbb\") pod \"mariadb-operator-controller-manager-88c7-tgx7p\" (UID: \"7c2eb266-7f51-45b9-8095-89d219bb59f0\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p" Oct 02 09:56:01 crc kubenswrapper[4771]: I1002 09:56:01.967804 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5njdl\" (UniqueName: \"kubernetes.io/projected/04a0b3f5-6e78-477a-8bd6-9250ba41eabf-kube-api-access-5njdl\") pod \"manila-operator-controller-manager-6d68dbc695-4dcd9\" (UID: \"04a0b3f5-6e78-477a-8bd6-9250ba41eabf\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.016413 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp"] Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.027685 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.046247 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-67mw2" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.057518 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj"] Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.063923 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwggn\" (UniqueName: \"kubernetes.io/projected/ac8e19b1-242e-4785-96ee-4c72ee975cce-kube-api-access-wwggn\") pod \"ovn-operator-controller-manager-9976ff44c-gs7d5\" (UID: \"ac8e19b1-242e-4785-96ee-4c72ee975cce\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.064077 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ffc86aad-8b08-40c7-85f9-fb498bbf56b2-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-lpb8d\" (UID: \"ffc86aad-8b08-40c7-85f9-fb498bbf56b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.064189 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc7jg\" (UniqueName: \"kubernetes.io/projected/9b273823-7cb9-4c85-b5cc-54a52a5d04b8-kube-api-access-jc7jg\") pod \"nova-operator-controller-manager-64cd67b5cb-mrfv5\" (UID: \"9b273823-7cb9-4c85-b5cc-54a52a5d04b8\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.064339 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wf88\" (UniqueName: \"kubernetes.io/projected/ffc86aad-8b08-40c7-85f9-fb498bbf56b2-kube-api-access-6wf88\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-lpb8d\" (UID: \"ffc86aad-8b08-40c7-85f9-fb498bbf56b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.064480 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lhp7\" (UniqueName: \"kubernetes.io/projected/4a399512-1d85-4f61-b8d4-800a57c33f4d-kube-api-access-7lhp7\") pod \"neutron-operator-controller-manager-849d5b9b84-zwjcd\" (UID: \"4a399512-1d85-4f61-b8d4-800a57c33f4d\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.064535 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qfdj\" (UniqueName: \"kubernetes.io/projected/4442fde9-3b77-4013-a6c2-87cfeca2b299-kube-api-access-2qfdj\") pod \"octavia-operator-controller-manager-7b787867f4-4f4vk\" (UID: \"4442fde9-3b77-4013-a6c2-87cfeca2b299\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.092363 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.097402 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-gxdt8" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.100392 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.110679 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lhp7\" (UniqueName: \"kubernetes.io/projected/4a399512-1d85-4f61-b8d4-800a57c33f4d-kube-api-access-7lhp7\") pod \"neutron-operator-controller-manager-849d5b9b84-zwjcd\" (UID: \"4a399512-1d85-4f61-b8d4-800a57c33f4d\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.124780 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz"] Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.142948 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.125920 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.149872 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.150035 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.254054 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-vzmgk" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.179145 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsrv4\" (UniqueName: \"kubernetes.io/projected/57aab528-153f-4c43-95d6-0c076d2071df-kube-api-access-qsrv4\") pod \"placement-operator-controller-manager-589c58c6c-l76nz\" (UID: \"57aab528-153f-4c43-95d6-0c076d2071df\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.264269 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwggn\" (UniqueName: \"kubernetes.io/projected/ac8e19b1-242e-4785-96ee-4c72ee975cce-kube-api-access-wwggn\") pod \"ovn-operator-controller-manager-9976ff44c-gs7d5\" (UID: \"ac8e19b1-242e-4785-96ee-4c72ee975cce\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.264379 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ffc86aad-8b08-40c7-85f9-fb498bbf56b2-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-lpb8d\" (UID: \"ffc86aad-8b08-40c7-85f9-fb498bbf56b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.264453 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc7jg\" (UniqueName: \"kubernetes.io/projected/9b273823-7cb9-4c85-b5cc-54a52a5d04b8-kube-api-access-jc7jg\") pod \"nova-operator-controller-manager-64cd67b5cb-mrfv5\" (UID: \"9b273823-7cb9-4c85-b5cc-54a52a5d04b8\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.264541 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2jdf\" (UniqueName: \"kubernetes.io/projected/a8a3258c-981f-4f61-bb08-42b5065a9067-kube-api-access-f2jdf\") pod \"telemetry-operator-controller-manager-d4856d54-q8sbj\" (UID: \"a8a3258c-981f-4f61-bb08-42b5065a9067\") " pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.188315 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.264944 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k9kp\" (UniqueName: \"kubernetes.io/projected/7f64d01d-a2f3-4e96-9f81-13cafb80225e-kube-api-access-6k9kp\") pod \"swift-operator-controller-manager-84d6b4b759-hxpcp\" (UID: \"7f64d01d-a2f3-4e96-9f81-13cafb80225e\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.265005 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wf88\" (UniqueName: \"kubernetes.io/projected/ffc86aad-8b08-40c7-85f9-fb498bbf56b2-kube-api-access-6wf88\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-lpb8d\" (UID: \"ffc86aad-8b08-40c7-85f9-fb498bbf56b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.205317 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp"] Oct 02 09:56:02 crc kubenswrapper[4771]: E1002 09:56:02.265354 4771 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 09:56:02 crc kubenswrapper[4771]: E1002 09:56:02.268934 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ffc86aad-8b08-40c7-85f9-fb498bbf56b2-cert podName:ffc86aad-8b08-40c7-85f9-fb498bbf56b2 nodeName:}" failed. No retries permitted until 2025-10-02 09:56:02.76890732 +0000 UTC m=+1150.416592467 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ffc86aad-8b08-40c7-85f9-fb498bbf56b2-cert") pod "openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" (UID: "ffc86aad-8b08-40c7-85f9-fb498bbf56b2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.265142 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qfdj\" (UniqueName: \"kubernetes.io/projected/4442fde9-3b77-4013-a6c2-87cfeca2b299-kube-api-access-2qfdj\") pod \"octavia-operator-controller-manager-7b787867f4-4f4vk\" (UID: \"4442fde9-3b77-4013-a6c2-87cfeca2b299\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.327934 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc7jg\" (UniqueName: \"kubernetes.io/projected/9b273823-7cb9-4c85-b5cc-54a52a5d04b8-kube-api-access-jc7jg\") pod \"nova-operator-controller-manager-64cd67b5cb-mrfv5\" (UID: \"9b273823-7cb9-4c85-b5cc-54a52a5d04b8\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.343630 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwggn\" (UniqueName: \"kubernetes.io/projected/ac8e19b1-242e-4785-96ee-4c72ee975cce-kube-api-access-wwggn\") pod \"ovn-operator-controller-manager-9976ff44c-gs7d5\" (UID: \"ac8e19b1-242e-4785-96ee-4c72ee975cce\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.343720 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-59m2p"] Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.345742 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.345761 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wf88\" (UniqueName: \"kubernetes.io/projected/ffc86aad-8b08-40c7-85f9-fb498bbf56b2-kube-api-access-6wf88\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-lpb8d\" (UID: \"ffc86aad-8b08-40c7-85f9-fb498bbf56b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.347768 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qfdj\" (UniqueName: \"kubernetes.io/projected/4442fde9-3b77-4013-a6c2-87cfeca2b299-kube-api-access-2qfdj\") pod \"octavia-operator-controller-manager-7b787867f4-4f4vk\" (UID: \"4442fde9-3b77-4013-a6c2-87cfeca2b299\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.353532 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-mr2sw" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.366543 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj"] Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.390470 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k9kp\" (UniqueName: \"kubernetes.io/projected/7f64d01d-a2f3-4e96-9f81-13cafb80225e-kube-api-access-6k9kp\") pod \"swift-operator-controller-manager-84d6b4b759-hxpcp\" (UID: \"7f64d01d-a2f3-4e96-9f81-13cafb80225e\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.390591 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e0df3275-79e0-49b8-93d5-585e2667a5c4-cert\") pod \"infra-operator-controller-manager-9d6c5db85-99mgr\" (UID: \"e0df3275-79e0-49b8-93d5-585e2667a5c4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.390721 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjwqb\" (UniqueName: \"kubernetes.io/projected/5c5a6559-fbef-41e7-bd03-71e40ef7ff7b-kube-api-access-rjwqb\") pod \"test-operator-controller-manager-85777745bb-59m2p\" (UID: \"5c5a6559-fbef-41e7-bd03-71e40ef7ff7b\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.390749 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsrv4\" (UniqueName: \"kubernetes.io/projected/57aab528-153f-4c43-95d6-0c076d2071df-kube-api-access-qsrv4\") pod \"placement-operator-controller-manager-589c58c6c-l76nz\" (UID: \"57aab528-153f-4c43-95d6-0c076d2071df\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.390918 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2jdf\" (UniqueName: \"kubernetes.io/projected/a8a3258c-981f-4f61-bb08-42b5065a9067-kube-api-access-f2jdf\") pod \"telemetry-operator-controller-manager-d4856d54-q8sbj\" (UID: \"a8a3258c-981f-4f61-bb08-42b5065a9067\") " pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.399764 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e0df3275-79e0-49b8-93d5-585e2667a5c4-cert\") pod \"infra-operator-controller-manager-9d6c5db85-99mgr\" (UID: \"e0df3275-79e0-49b8-93d5-585e2667a5c4\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.413993 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.463247 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2jdf\" (UniqueName: \"kubernetes.io/projected/a8a3258c-981f-4f61-bb08-42b5065a9067-kube-api-access-f2jdf\") pod \"telemetry-operator-controller-manager-d4856d54-q8sbj\" (UID: \"a8a3258c-981f-4f61-bb08-42b5065a9067\") " pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.467492 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.471829 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsrv4\" (UniqueName: \"kubernetes.io/projected/57aab528-153f-4c43-95d6-0c076d2071df-kube-api-access-qsrv4\") pod \"placement-operator-controller-manager-589c58c6c-l76nz\" (UID: \"57aab528-153f-4c43-95d6-0c076d2071df\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.485232 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz"] Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.509572 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.510422 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjwqb\" (UniqueName: \"kubernetes.io/projected/5c5a6559-fbef-41e7-bd03-71e40ef7ff7b-kube-api-access-rjwqb\") pod \"test-operator-controller-manager-85777745bb-59m2p\" (UID: \"5c5a6559-fbef-41e7-bd03-71e40ef7ff7b\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.517737 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6"] Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.519324 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.528654 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-67v7c" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.529639 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.541090 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k9kp\" (UniqueName: \"kubernetes.io/projected/7f64d01d-a2f3-4e96-9f81-13cafb80225e-kube-api-access-6k9kp\") pod \"swift-operator-controller-manager-84d6b4b759-hxpcp\" (UID: \"7f64d01d-a2f3-4e96-9f81-13cafb80225e\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.577103 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.607060 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjwqb\" (UniqueName: \"kubernetes.io/projected/5c5a6559-fbef-41e7-bd03-71e40ef7ff7b-kube-api-access-rjwqb\") pod \"test-operator-controller-manager-85777745bb-59m2p\" (UID: \"5c5a6559-fbef-41e7-bd03-71e40ef7ff7b\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.613189 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k289h\" (UniqueName: \"kubernetes.io/projected/f43ba614-8bea-47db-be7d-10c7a4e9ddbb-kube-api-access-k289h\") pod \"watcher-operator-controller-manager-6b9957f54f-wmrf6\" (UID: \"f43ba614-8bea-47db-be7d-10c7a4e9ddbb\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.629255 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6"] Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.658002 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.663326 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-59m2p"] Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.703450 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.724738 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k289h\" (UniqueName: \"kubernetes.io/projected/f43ba614-8bea-47db-be7d-10c7a4e9ddbb-kube-api-access-k289h\") pod \"watcher-operator-controller-manager-6b9957f54f-wmrf6\" (UID: \"f43ba614-8bea-47db-be7d-10c7a4e9ddbb\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.748722 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.787174 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k289h\" (UniqueName: \"kubernetes.io/projected/f43ba614-8bea-47db-be7d-10c7a4e9ddbb-kube-api-access-k289h\") pod \"watcher-operator-controller-manager-6b9957f54f-wmrf6\" (UID: \"f43ba614-8bea-47db-be7d-10c7a4e9ddbb\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.807720 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5"] Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.809399 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.821617 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.821784 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.821922 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-2zh2s" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.826570 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ffc86aad-8b08-40c7-85f9-fb498bbf56b2-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-lpb8d\" (UID: \"ffc86aad-8b08-40c7-85f9-fb498bbf56b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.834563 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5"] Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.842199 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ffc86aad-8b08-40c7-85f9-fb498bbf56b2-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-lpb8d\" (UID: \"ffc86aad-8b08-40c7-85f9-fb498bbf56b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.901611 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l"] Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.902927 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.906643 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-q5rhr" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.910257 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l"] Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.929770 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkggw\" (UniqueName: \"kubernetes.io/projected/6925b4e5-a255-4f79-b833-8862a67dc0ea-kube-api-access-bkggw\") pod \"openstack-operator-controller-manager-6567974949-gkcw5\" (UID: \"6925b4e5-a255-4f79-b833-8862a67dc0ea\") " pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.930209 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6925b4e5-a255-4f79-b833-8862a67dc0ea-cert\") pod \"openstack-operator-controller-manager-6567974949-gkcw5\" (UID: \"6925b4e5-a255-4f79-b833-8862a67dc0ea\") " pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" Oct 02 09:56:02 crc kubenswrapper[4771]: I1002 09:56:02.941608 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.032269 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlqmq\" (UniqueName: \"kubernetes.io/projected/ca4a791e-d59a-47c3-8f39-8d552c463c64-kube-api-access-mlqmq\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l\" (UID: \"ca4a791e-d59a-47c3-8f39-8d552c463c64\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l" Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.032344 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6925b4e5-a255-4f79-b833-8862a67dc0ea-cert\") pod \"openstack-operator-controller-manager-6567974949-gkcw5\" (UID: \"6925b4e5-a255-4f79-b833-8862a67dc0ea\") " pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.032411 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkggw\" (UniqueName: \"kubernetes.io/projected/6925b4e5-a255-4f79-b833-8862a67dc0ea-kube-api-access-bkggw\") pod \"openstack-operator-controller-manager-6567974949-gkcw5\" (UID: \"6925b4e5-a255-4f79-b833-8862a67dc0ea\") " pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.062625 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6925b4e5-a255-4f79-b833-8862a67dc0ea-cert\") pod \"openstack-operator-controller-manager-6567974949-gkcw5\" (UID: \"6925b4e5-a255-4f79-b833-8862a67dc0ea\") " pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.077771 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkggw\" (UniqueName: \"kubernetes.io/projected/6925b4e5-a255-4f79-b833-8862a67dc0ea-kube-api-access-bkggw\") pod \"openstack-operator-controller-manager-6567974949-gkcw5\" (UID: \"6925b4e5-a255-4f79-b833-8862a67dc0ea\") " pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.134501 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlqmq\" (UniqueName: \"kubernetes.io/projected/ca4a791e-d59a-47c3-8f39-8d552c463c64-kube-api-access-mlqmq\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l\" (UID: \"ca4a791e-d59a-47c3-8f39-8d552c463c64\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l" Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.161578 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlqmq\" (UniqueName: \"kubernetes.io/projected/ca4a791e-d59a-47c3-8f39-8d552c463c64-kube-api-access-mlqmq\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l\" (UID: \"ca4a791e-d59a-47c3-8f39-8d552c463c64\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l" Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.253067 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.308043 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l" Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.717173 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw"] Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.921245 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd"] Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.933600 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h"] Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.940822 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj"] Oct 02 09:56:03 crc kubenswrapper[4771]: I1002 09:56:03.947590 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw"] Oct 02 09:56:04 crc kubenswrapper[4771]: I1002 09:56:04.477505 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj" event={"ID":"8e8d717b-41cc-4dd7-b9e3-4d06ac36cc54","Type":"ContainerStarted","Data":"66cc64ede1ee038d89b9dbf5bc82056561252224659595da0c3c1fc06a63c552"} Oct 02 09:56:04 crc kubenswrapper[4771]: I1002 09:56:04.480309 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw" event={"ID":"f0638dc6-d014-4a78-bcfa-eb4e5cea5caf","Type":"ContainerStarted","Data":"fef526595727d4ae72a4891a87efca4b6c222ac00b2b40a1ea8c4214aae60a36"} Oct 02 09:56:04 crc kubenswrapper[4771]: I1002 09:56:04.485045 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" event={"ID":"6dc77080-a5ab-4db4-b150-2573589da332","Type":"ContainerStarted","Data":"4b4e27de5d3763445c9e63b99307ded1579be38466d08187fac0e8a780b724be"} Oct 02 09:56:04 crc kubenswrapper[4771]: I1002 09:56:04.486467 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h" event={"ID":"4da5dda4-f0da-453d-ba80-a2fc95035688","Type":"ContainerStarted","Data":"23b408d11206c1e1424b95c9fa73a316ab3a005e1a530905d566d5f3015b917e"} Oct 02 09:56:04 crc kubenswrapper[4771]: I1002 09:56:04.487687 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw" event={"ID":"5fcc6b1c-2163-454b-9172-16852b9f8966","Type":"ContainerStarted","Data":"b7a8546e20a9ac391d2def6cba859ba3921f1f658444ef6d35ee380554870a7c"} Oct 02 09:56:04 crc kubenswrapper[4771]: I1002 09:56:04.966347 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p"] Oct 02 09:56:04 crc kubenswrapper[4771]: I1002 09:56:04.997037 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz"] Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.067824 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk"] Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.085276 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l"] Oct 02 09:56:05 crc kubenswrapper[4771]: W1002 09:56:05.094062 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f236f0a_744a_4f98_90e9_7c78ead31ddc.slice/crio-33aaacc54946e1445126d03cadd58b5cd7c860d9cf329523a8b9e90f4e505f92 WatchSource:0}: Error finding container 33aaacc54946e1445126d03cadd58b5cd7c860d9cf329523a8b9e90f4e505f92: Status 404 returned error can't find the container with id 33aaacc54946e1445126d03cadd58b5cd7c860d9cf329523a8b9e90f4e505f92 Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.104787 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv"] Oct 02 09:56:05 crc kubenswrapper[4771]: W1002 09:56:05.108318 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57aab528_153f_4c43_95d6_0c076d2071df.slice/crio-6297ce8337c09872df2647c199975e919ea0616116b358f34b9d7378c48bf467 WatchSource:0}: Error finding container 6297ce8337c09872df2647c199975e919ea0616116b358f34b9d7378c48bf467: Status 404 returned error can't find the container with id 6297ce8337c09872df2647c199975e919ea0616116b358f34b9d7378c48bf467 Oct 02 09:56:05 crc kubenswrapper[4771]: W1002 09:56:05.113729 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b273823_7cb9_4c85_b5cc_54a52a5d04b8.slice/crio-da06ab43820aae13ce3da0a7065ec8a38b358b200f3b6b13938e411b8a0bf656 WatchSource:0}: Error finding container da06ab43820aae13ce3da0a7065ec8a38b358b200f3b6b13938e411b8a0bf656: Status 404 returned error can't find the container with id da06ab43820aae13ce3da0a7065ec8a38b358b200f3b6b13938e411b8a0bf656 Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.121560 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5"] Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.134717 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9"] Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.162018 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr"] Oct 02 09:56:05 crc kubenswrapper[4771]: W1002 09:56:05.194494 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4442fde9_3b77_4013_a6c2_87cfeca2b299.slice/crio-d040ba0652c0f180d7664e17e024b8c588836de2fdca1eb040da6b9ef5ddb46c WatchSource:0}: Error finding container d040ba0652c0f180d7664e17e024b8c588836de2fdca1eb040da6b9ef5ddb46c: Status 404 returned error can't find the container with id d040ba0652c0f180d7664e17e024b8c588836de2fdca1eb040da6b9ef5ddb46c Oct 02 09:56:05 crc kubenswrapper[4771]: E1002 09:56:05.264506 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rjwqb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-85777745bb-59m2p_openstack-operators(5c5a6559-fbef-41e7-bd03-71e40ef7ff7b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 09:56:05 crc kubenswrapper[4771]: E1002 09:56:05.270238 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wvf25,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-5bd55b4bff-jg6rv_openstack-operators(1f236f0a-744a-4f98-90e9-7c78ead31ddc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.278852 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd"] Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.306018 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5"] Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.346670 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj"] Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.361179 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-59m2p"] Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.384254 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756"] Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.430259 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6"] Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.493702 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp"] Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.511244 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5"] Oct 02 09:56:05 crc kubenswrapper[4771]: E1002 09:56:05.519684 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6k9kp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-84d6b4b759-hxpcp_openstack-operators(7f64d01d-a2f3-4e96-9f81-13cafb80225e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.520465 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd" event={"ID":"4a399512-1d85-4f61-b8d4-800a57c33f4d","Type":"ContainerStarted","Data":"94a55b4b8d2de5b6332e80abebb6831013ec926124e6d5148b60f2856a7066e3"} Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.521060 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d"] Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.525345 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" event={"ID":"04a0b3f5-6e78-477a-8bd6-9250ba41eabf","Type":"ContainerStarted","Data":"dcca7f06dec7b4e067523d539dda8c21755c212033d672e08e3f9db945e18183"} Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.546962 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l"] Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.565891 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz" event={"ID":"57aab528-153f-4c43-95d6-0c076d2071df","Type":"ContainerStarted","Data":"6297ce8337c09872df2647c199975e919ea0616116b358f34b9d7378c48bf467"} Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.596360 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" event={"ID":"e0df3275-79e0-49b8-93d5-585e2667a5c4","Type":"ContainerStarted","Data":"ef7e1b5393e11c62624311590a89ca38f2fa61c842f897f80b02680a6333dbbb"} Oct 02 09:56:05 crc kubenswrapper[4771]: E1002 09:56:05.599168 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mlqmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l_openstack-operators(ca4a791e-d59a-47c3-8f39-8d552c463c64): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 09:56:05 crc kubenswrapper[4771]: E1002 09:56:05.599469 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6wf88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-5869cb545-lpb8d_openstack-operators(ffc86aad-8b08-40c7-85f9-fb498bbf56b2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 09:56:05 crc kubenswrapper[4771]: E1002 09:56:05.601988 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l" podUID="ca4a791e-d59a-47c3-8f39-8d552c463c64" Oct 02 09:56:05 crc kubenswrapper[4771]: E1002 09:56:05.602232 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" podUID="5c5a6559-fbef-41e7-bd03-71e40ef7ff7b" Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.603552 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk" event={"ID":"4442fde9-3b77-4013-a6c2-87cfeca2b299","Type":"ContainerStarted","Data":"d040ba0652c0f180d7664e17e024b8c588836de2fdca1eb040da6b9ef5ddb46c"} Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.607180 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l" event={"ID":"7e625416-21a2-45e6-aec5-cf04c408e65e","Type":"ContainerStarted","Data":"cba018f37e252f35d00f7a1027fe922abd464cb7a6e7b599a55883c9c69f2590"} Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.623107 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" event={"ID":"5c5a6559-fbef-41e7-bd03-71e40ef7ff7b","Type":"ContainerStarted","Data":"7989091952fc4507c327218472238b93839d2657c0c1864ecdfbf08d1240b88f"} Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.631697 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p" event={"ID":"7c2eb266-7f51-45b9-8095-89d219bb59f0","Type":"ContainerStarted","Data":"608d335e4e99d8ef84286c351c6a3505e78b281fa011ae3ee28426db02255911"} Oct 02 09:56:05 crc kubenswrapper[4771]: E1002 09:56:05.638561 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" podUID="1f236f0a-744a-4f98-90e9-7c78ead31ddc" Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.648467 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" event={"ID":"a8a3258c-981f-4f61-bb08-42b5065a9067","Type":"ContainerStarted","Data":"40b4b25f741df09cb2e111d3aca8fce2248528e4dbca411cf05de4fc7855cf00"} Oct 02 09:56:05 crc kubenswrapper[4771]: E1002 09:56:05.655842 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" podUID="5c5a6559-fbef-41e7-bd03-71e40ef7ff7b" Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.656947 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6" event={"ID":"f43ba614-8bea-47db-be7d-10c7a4e9ddbb","Type":"ContainerStarted","Data":"797cfccd23044ba58360a01f6b9c414e81309c017a3fb8e5f94563dfbeb59d4b"} Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.782905 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756" event={"ID":"9804f0f6-0273-4b09-830f-f082cf718803","Type":"ContainerStarted","Data":"f201a02d43e79f58b06308e7dba98cad4121b80a40cb13b7e956beea614667cb"} Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.782949 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5" event={"ID":"9b273823-7cb9-4c85-b5cc-54a52a5d04b8","Type":"ContainerStarted","Data":"da06ab43820aae13ce3da0a7065ec8a38b358b200f3b6b13938e411b8a0bf656"} Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.782962 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5" event={"ID":"ac8e19b1-242e-4785-96ee-4c72ee975cce","Type":"ContainerStarted","Data":"70c792a476abd454cb454eb0b4204d6a4c598fd5861209c59be5dd9b288cc744"} Oct 02 09:56:05 crc kubenswrapper[4771]: I1002 09:56:05.795203 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" event={"ID":"1f236f0a-744a-4f98-90e9-7c78ead31ddc","Type":"ContainerStarted","Data":"33aaacc54946e1445126d03cadd58b5cd7c860d9cf329523a8b9e90f4e505f92"} Oct 02 09:56:05 crc kubenswrapper[4771]: E1002 09:56:05.797773 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" podUID="1f236f0a-744a-4f98-90e9-7c78ead31ddc" Oct 02 09:56:06 crc kubenswrapper[4771]: E1002 09:56:06.182940 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" podUID="7f64d01d-a2f3-4e96-9f81-13cafb80225e" Oct 02 09:56:06 crc kubenswrapper[4771]: E1002 09:56:06.186405 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" podUID="ffc86aad-8b08-40c7-85f9-fb498bbf56b2" Oct 02 09:56:06 crc kubenswrapper[4771]: I1002 09:56:06.839115 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l" event={"ID":"ca4a791e-d59a-47c3-8f39-8d552c463c64","Type":"ContainerStarted","Data":"2c887c8f59432fa71faec0e54a99db8c641123c559a3963376b4e3d01cdb2a9c"} Oct 02 09:56:06 crc kubenswrapper[4771]: E1002 09:56:06.844927 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l" podUID="ca4a791e-d59a-47c3-8f39-8d552c463c64" Oct 02 09:56:06 crc kubenswrapper[4771]: I1002 09:56:06.854278 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" event={"ID":"1f236f0a-744a-4f98-90e9-7c78ead31ddc","Type":"ContainerStarted","Data":"211007d64ca5f70974867054bacc1307e17e9841f6080c68fbd99e7e8ada1540"} Oct 02 09:56:06 crc kubenswrapper[4771]: E1002 09:56:06.860125 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" podUID="1f236f0a-744a-4f98-90e9-7c78ead31ddc" Oct 02 09:56:06 crc kubenswrapper[4771]: I1002 09:56:06.877926 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" event={"ID":"5c5a6559-fbef-41e7-bd03-71e40ef7ff7b","Type":"ContainerStarted","Data":"9ffc48a07bf17dd391f5b7273b1eb00c1186cab5fd0fa49fcc8274f27fadadb2"} Oct 02 09:56:06 crc kubenswrapper[4771]: E1002 09:56:06.883429 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" podUID="5c5a6559-fbef-41e7-bd03-71e40ef7ff7b" Oct 02 09:56:06 crc kubenswrapper[4771]: I1002 09:56:06.896731 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" event={"ID":"ffc86aad-8b08-40c7-85f9-fb498bbf56b2","Type":"ContainerStarted","Data":"eb7e968390bc703eb61c2ff3809763026b10bd7537308a115646c7c7fa9a530f"} Oct 02 09:56:06 crc kubenswrapper[4771]: I1002 09:56:06.896790 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" event={"ID":"ffc86aad-8b08-40c7-85f9-fb498bbf56b2","Type":"ContainerStarted","Data":"6c1434b3caee7228d89793d0376c9d5fe6d66c3208b62a6abc60e9356d3f53f5"} Oct 02 09:56:06 crc kubenswrapper[4771]: E1002 09:56:06.902027 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" podUID="ffc86aad-8b08-40c7-85f9-fb498bbf56b2" Oct 02 09:56:06 crc kubenswrapper[4771]: I1002 09:56:06.916348 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" event={"ID":"7f64d01d-a2f3-4e96-9f81-13cafb80225e","Type":"ContainerStarted","Data":"1f3ed08a0753190d6f8a139130e7da8405c13cebc46fe33f3c5a98e7fb9f7e7e"} Oct 02 09:56:06 crc kubenswrapper[4771]: I1002 09:56:06.916396 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" event={"ID":"7f64d01d-a2f3-4e96-9f81-13cafb80225e","Type":"ContainerStarted","Data":"8b97511778dec96c8efbbe3c4a865ad5fbe14e12d77cca7f797328c078c9d49e"} Oct 02 09:56:06 crc kubenswrapper[4771]: E1002 09:56:06.923735 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" podUID="7f64d01d-a2f3-4e96-9f81-13cafb80225e" Oct 02 09:56:06 crc kubenswrapper[4771]: I1002 09:56:06.924621 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" event={"ID":"6925b4e5-a255-4f79-b833-8862a67dc0ea","Type":"ContainerStarted","Data":"32af673787e8f99dd844475dc67ea22eeea7858b755acfb417a6ed1953265865"} Oct 02 09:56:06 crc kubenswrapper[4771]: I1002 09:56:06.924672 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" event={"ID":"6925b4e5-a255-4f79-b833-8862a67dc0ea","Type":"ContainerStarted","Data":"237b3b52a83000a25f445d7010638643ce271e36f7cb85fb8edabcabaf98f7f7"} Oct 02 09:56:06 crc kubenswrapper[4771]: I1002 09:56:06.924683 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" event={"ID":"6925b4e5-a255-4f79-b833-8862a67dc0ea","Type":"ContainerStarted","Data":"eefe51b701f915e335148ded59645ff4f5cd62dd464f3c10a11e535c96c68648"} Oct 02 09:56:07 crc kubenswrapper[4771]: E1002 09:56:07.939554 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l" podUID="ca4a791e-d59a-47c3-8f39-8d552c463c64" Oct 02 09:56:07 crc kubenswrapper[4771]: E1002 09:56:07.940147 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" podUID="7f64d01d-a2f3-4e96-9f81-13cafb80225e" Oct 02 09:56:07 crc kubenswrapper[4771]: E1002 09:56:07.940331 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" podUID="ffc86aad-8b08-40c7-85f9-fb498bbf56b2" Oct 02 09:56:07 crc kubenswrapper[4771]: E1002 09:56:07.941115 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" podUID="1f236f0a-744a-4f98-90e9-7c78ead31ddc" Oct 02 09:56:07 crc kubenswrapper[4771]: E1002 09:56:07.944690 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" podUID="5c5a6559-fbef-41e7-bd03-71e40ef7ff7b" Oct 02 09:56:08 crc kubenswrapper[4771]: I1002 09:56:08.072026 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" podStartSLOduration=6.071991689 podStartE2EDuration="6.071991689s" podCreationTimestamp="2025-10-02 09:56:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:56:08.058837373 +0000 UTC m=+1155.706522440" watchObservedRunningTime="2025-10-02 09:56:08.071991689 +0000 UTC m=+1155.719676756" Oct 02 09:56:12 crc kubenswrapper[4771]: I1002 09:56:12.146644 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:56:12 crc kubenswrapper[4771]: I1002 09:56:12.147293 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:56:12 crc kubenswrapper[4771]: I1002 09:56:12.147354 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:56:12 crc kubenswrapper[4771]: I1002 09:56:12.148344 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9385c2320866e6d26f67de876609b5bcc62a83b76b282bedf5fb7cfa0fec4a42"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:56:12 crc kubenswrapper[4771]: I1002 09:56:12.148416 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://9385c2320866e6d26f67de876609b5bcc62a83b76b282bedf5fb7cfa0fec4a42" gracePeriod=600 Oct 02 09:56:13 crc kubenswrapper[4771]: I1002 09:56:13.008267 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="9385c2320866e6d26f67de876609b5bcc62a83b76b282bedf5fb7cfa0fec4a42" exitCode=0 Oct 02 09:56:13 crc kubenswrapper[4771]: I1002 09:56:13.008339 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"9385c2320866e6d26f67de876609b5bcc62a83b76b282bedf5fb7cfa0fec4a42"} Oct 02 09:56:13 crc kubenswrapper[4771]: I1002 09:56:13.008633 4771 scope.go:117] "RemoveContainer" containerID="e881357965830ad6d553dbf5d22ee7a11da1ee52ace8a1671edcdf44ed33436b" Oct 02 09:56:13 crc kubenswrapper[4771]: I1002 09:56:13.254228 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" Oct 02 09:56:13 crc kubenswrapper[4771]: I1002 09:56:13.259661 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6567974949-gkcw5" Oct 02 09:56:17 crc kubenswrapper[4771]: E1002 09:56:17.925255 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397" Oct 02 09:56:17 crc kubenswrapper[4771]: E1002 09:56:17.926116 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xfgg4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-9f4696d94-lp8xj_openstack-operators(8e8d717b-41cc-4dd7-b9e3-4d06ac36cc54): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:56:19 crc kubenswrapper[4771]: E1002 09:56:19.234976 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:15d7b5a365350a831ca59d984df67fadeccf89d599e487a7597b105afb82ce4a" Oct 02 09:56:19 crc kubenswrapper[4771]: E1002 09:56:19.235704 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:15d7b5a365350a831ca59d984df67fadeccf89d599e487a7597b105afb82ce4a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-67zbb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-88c7-tgx7p_openstack-operators(7c2eb266-7f51-45b9-8095-89d219bb59f0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:56:20 crc kubenswrapper[4771]: E1002 09:56:20.825457 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:bb39758cc8cd0d2cd02841dc81b53fd88647e2db15ee16cdd8c44d4098a942fd" Oct 02 09:56:20 crc kubenswrapper[4771]: E1002 09:56:20.826072 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:bb39758cc8cd0d2cd02841dc81b53fd88647e2db15ee16cdd8c44d4098a942fd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6pjpv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-6ff8b75857-smc8h_openstack-operators(4da5dda4-f0da-453d-ba80-a2fc95035688): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:56:21 crc kubenswrapper[4771]: E1002 09:56:21.564185 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:acdeebaa51f962066f42f38b6c2d34a62fc6a24f58f9ee63d61b1e0cafbb29f8" Oct 02 09:56:21 crc kubenswrapper[4771]: E1002 09:56:21.564672 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:acdeebaa51f962066f42f38b6c2d34a62fc6a24f58f9ee63d61b1e0cafbb29f8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7lhp7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-849d5b9b84-zwjcd_openstack-operators(4a399512-1d85-4f61-b8d4-800a57c33f4d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:56:26 crc kubenswrapper[4771]: E1002 09:56:26.385677 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2" Oct 02 09:56:26 crc kubenswrapper[4771]: E1002 09:56:26.386439 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qsrv4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-589c58c6c-l76nz_openstack-operators(57aab528-153f-4c43-95d6-0c076d2071df): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:56:27 crc kubenswrapper[4771]: E1002 09:56:27.130611 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:917e6dcc519277c46e42898bc9f0f066790fa7b9633fcde668cc8a68a547c13c" Oct 02 09:56:27 crc kubenswrapper[4771]: E1002 09:56:27.130830 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:917e6dcc519277c46e42898bc9f0f066790fa7b9633fcde668cc8a68a547c13c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vfzcj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5d889d78cf-hh756_openstack-operators(9804f0f6-0273-4b09-830f-f082cf718803): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:56:27 crc kubenswrapper[4771]: E1002 09:56:27.481638 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9" Oct 02 09:56:27 crc kubenswrapper[4771]: E1002 09:56:27.482908 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2qfdj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-7b787867f4-4f4vk_openstack-operators(4442fde9-3b77-4013-a6c2-87cfeca2b299): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:56:28 crc kubenswrapper[4771]: E1002 09:56:28.177363 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f" Oct 02 09:56:28 crc kubenswrapper[4771]: E1002 09:56:28.177679 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jc7jg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-64cd67b5cb-mrfv5_openstack-operators(9b273823-7cb9-4c85-b5cc-54a52a5d04b8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:56:30 crc kubenswrapper[4771]: E1002 09:56:30.734216 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884" Oct 02 09:56:30 crc kubenswrapper[4771]: E1002 09:56:30.734751 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5njdl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6d68dbc695-4dcd9_openstack-operators(04a0b3f5-6e78-477a-8bd6-9250ba41eabf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:56:34 crc kubenswrapper[4771]: E1002 09:56:34.015681 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:21792a2317c0a55e40b2a02a7d5d4682b76538ed2a2e0633199aa395e60ecc72" Oct 02 09:56:34 crc kubenswrapper[4771]: E1002 09:56:34.016614 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:21792a2317c0a55e40b2a02a7d5d4682b76538ed2a2e0633199aa395e60ecc72,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vhv6v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-84958c4d49-pnqnd_openstack-operators(6dc77080-a5ab-4db4-b150-2573589da332): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:56:34 crc kubenswrapper[4771]: E1002 09:56:34.215055 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/openstack-k8s-operators/telemetry-operator:114f1d2f7a3b244c57adf7a27aa58cc483e58158" Oct 02 09:56:34 crc kubenswrapper[4771]: E1002 09:56:34.215145 4771 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.129.56.48:5001/openstack-k8s-operators/telemetry-operator:114f1d2f7a3b244c57adf7a27aa58cc483e58158" Oct 02 09:56:34 crc kubenswrapper[4771]: E1002 09:56:34.215340 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.129.56.48:5001/openstack-k8s-operators/telemetry-operator:114f1d2f7a3b244c57adf7a27aa58cc483e58158,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f2jdf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-d4856d54-q8sbj_openstack-operators(a8a3258c-981f-4f61-bb08-42b5065a9067): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:56:37 crc kubenswrapper[4771]: E1002 09:56:37.202662 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756" podUID="9804f0f6-0273-4b09-830f-f082cf718803" Oct 02 09:56:37 crc kubenswrapper[4771]: E1002 09:56:37.251244 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk" podUID="4442fde9-3b77-4013-a6c2-87cfeca2b299" Oct 02 09:56:37 crc kubenswrapper[4771]: E1002 09:56:37.251368 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" podUID="04a0b3f5-6e78-477a-8bd6-9250ba41eabf" Oct 02 09:56:37 crc kubenswrapper[4771]: E1002 09:56:37.257753 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj" podUID="8e8d717b-41cc-4dd7-b9e3-4d06ac36cc54" Oct 02 09:56:37 crc kubenswrapper[4771]: E1002 09:56:37.281652 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd" podUID="4a399512-1d85-4f61-b8d4-800a57c33f4d" Oct 02 09:56:37 crc kubenswrapper[4771]: I1002 09:56:37.283886 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756" event={"ID":"9804f0f6-0273-4b09-830f-f082cf718803","Type":"ContainerStarted","Data":"007db7bfa2e1f80b8d1474dd08b02f5811e32281b1be80dba8304ee6bbb91ac3"} Oct 02 09:56:37 crc kubenswrapper[4771]: I1002 09:56:37.289147 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6" event={"ID":"f43ba614-8bea-47db-be7d-10c7a4e9ddbb","Type":"ContainerStarted","Data":"0d8954a4bd9a9412273af3abb45e51b80d632d2a99bee6fde7f45b6c5bbf461e"} Oct 02 09:56:37 crc kubenswrapper[4771]: I1002 09:56:37.302446 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" event={"ID":"04a0b3f5-6e78-477a-8bd6-9250ba41eabf","Type":"ContainerStarted","Data":"546511ad6022d89817169fe90f39d87dcc6d34bc30c6cdaabfc40f7c581f4b24"} Oct 02 09:56:37 crc kubenswrapper[4771]: E1002 09:56:37.307547 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" podUID="04a0b3f5-6e78-477a-8bd6-9250ba41eabf" Oct 02 09:56:37 crc kubenswrapper[4771]: I1002 09:56:37.315040 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw" event={"ID":"5fcc6b1c-2163-454b-9172-16852b9f8966","Type":"ContainerStarted","Data":"ce1b5ed4d0789307e6e2cd105874646d8b8c0b25282f6aec46a3f242649f26b4"} Oct 02 09:56:37 crc kubenswrapper[4771]: I1002 09:56:37.322469 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj" event={"ID":"8e8d717b-41cc-4dd7-b9e3-4d06ac36cc54","Type":"ContainerStarted","Data":"349ebc92771f5c0ee8cf31fd946d3fd79311f9276ea7cab81c8e44eae4fb0789"} Oct 02 09:56:37 crc kubenswrapper[4771]: I1002 09:56:37.335410 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk" event={"ID":"4442fde9-3b77-4013-a6c2-87cfeca2b299","Type":"ContainerStarted","Data":"2711b1b2b200ab9be101202c0bddff163f3b335c41497809959c56d538c94729"} Oct 02 09:56:37 crc kubenswrapper[4771]: E1002 09:56:37.337652 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk" podUID="4442fde9-3b77-4013-a6c2-87cfeca2b299" Oct 02 09:56:37 crc kubenswrapper[4771]: E1002 09:56:37.371995 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5" podUID="9b273823-7cb9-4c85-b5cc-54a52a5d04b8" Oct 02 09:56:37 crc kubenswrapper[4771]: E1002 09:56:37.421909 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" podUID="6dc77080-a5ab-4db4-b150-2573589da332" Oct 02 09:56:37 crc kubenswrapper[4771]: E1002 09:56:37.452836 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz" podUID="57aab528-153f-4c43-95d6-0c076d2071df" Oct 02 09:56:37 crc kubenswrapper[4771]: E1002 09:56:37.536915 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h" podUID="4da5dda4-f0da-453d-ba80-a2fc95035688" Oct 02 09:56:37 crc kubenswrapper[4771]: E1002 09:56:37.623492 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p" podUID="7c2eb266-7f51-45b9-8095-89d219bb59f0" Oct 02 09:56:37 crc kubenswrapper[4771]: E1002 09:56:37.811480 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" podUID="a8a3258c-981f-4f61-bb08-42b5065a9067" Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.348152 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz" event={"ID":"57aab528-153f-4c43-95d6-0c076d2071df","Type":"ContainerStarted","Data":"5673f4aff1fa43579840260297d60746fadcb22697ce8c7e3bcd50f231be7017"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.350685 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" event={"ID":"6dc77080-a5ab-4db4-b150-2573589da332","Type":"ContainerStarted","Data":"ac0787ba16a2c9d91ca3e4d1b3ad6743fe8f7a9b32538a1d77bab7346ace4911"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.356842 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l" event={"ID":"7e625416-21a2-45e6-aec5-cf04c408e65e","Type":"ContainerStarted","Data":"369ea3e25ea61df568e6bcce8b010291d683eec23247fb810b708762d184add6"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.359591 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd" event={"ID":"4a399512-1d85-4f61-b8d4-800a57c33f4d","Type":"ContainerStarted","Data":"ca8863477af62d99e68eba995eecdb1d867304f5107d46dddd6300b041a89e85"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.362314 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" event={"ID":"5c5a6559-fbef-41e7-bd03-71e40ef7ff7b","Type":"ContainerStarted","Data":"c53cb7784eea5755cbfe6ec66478696001b6c031e5ce00db898178631abc1add"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.362615 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.364372 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p" event={"ID":"7c2eb266-7f51-45b9-8095-89d219bb59f0","Type":"ContainerStarted","Data":"30c072c0f239c67d03e59e990e7f3f16e83e4d96639bd9ebbfe4d82134536108"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.366046 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h" event={"ID":"4da5dda4-f0da-453d-ba80-a2fc95035688","Type":"ContainerStarted","Data":"92ca24387d3d8fdc932442ddecb104104274161ef3d4c4930a3c8bc231f48798"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.369983 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" event={"ID":"ffc86aad-8b08-40c7-85f9-fb498bbf56b2","Type":"ContainerStarted","Data":"d398479f745bc650b3f10ade7d8b86691fba3069b1237126d5e418d6ce8ad229"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.370449 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.372590 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"405cf9bc15cbb64122966859b742b139fd0128a50163288165936c501ca2e3e6"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.375420 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" event={"ID":"e0df3275-79e0-49b8-93d5-585e2667a5c4","Type":"ContainerStarted","Data":"2214272609bb0e0706f79e91134f0e420b842a7edeee0bf43474e07f412521a0"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.389804 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l" event={"ID":"ca4a791e-d59a-47c3-8f39-8d552c463c64","Type":"ContainerStarted","Data":"1f9a5cd5f932269c0b1d62d3b3631915ac19613acc5ae0fd346ce4530ffba07b"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.411779 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw" event={"ID":"f0638dc6-d014-4a78-bcfa-eb4e5cea5caf","Type":"ContainerStarted","Data":"cbbe6ea559b4e3ab374d8661a78cdcdd20ca54436f7bd92509d6d6b914da2701"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.414059 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5" event={"ID":"9b273823-7cb9-4c85-b5cc-54a52a5d04b8","Type":"ContainerStarted","Data":"1e0531d9fb092060f7c6889d0a752d0054cbef7d55a16a28ff1c6e87c2032e80"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.418216 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5" event={"ID":"ac8e19b1-242e-4785-96ee-4c72ee975cce","Type":"ContainerStarted","Data":"08b8220fddafa5afcfa69c6788085af66a1aa4268320dc6d0cbc26e45e812b0a"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.423095 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" event={"ID":"a8a3258c-981f-4f61-bb08-42b5065a9067","Type":"ContainerStarted","Data":"b5c2366639c27f456d5338516fbda1a68840f30b4599c3193d368ac26c29474e"} Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.427622 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" podStartSLOduration=5.934127574 podStartE2EDuration="37.427593475s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.264295894 +0000 UTC m=+1152.911980961" lastFinishedPulling="2025-10-02 09:56:36.757761795 +0000 UTC m=+1184.405446862" observedRunningTime="2025-10-02 09:56:38.421566681 +0000 UTC m=+1186.069251748" watchObservedRunningTime="2025-10-02 09:56:38.427593475 +0000 UTC m=+1186.075278542" Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.486097 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l" podStartSLOduration=5.212610834 podStartE2EDuration="36.48607495s" podCreationTimestamp="2025-10-02 09:56:02 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.598852689 +0000 UTC m=+1153.246537756" lastFinishedPulling="2025-10-02 09:56:36.872316805 +0000 UTC m=+1184.520001872" observedRunningTime="2025-10-02 09:56:38.471791885 +0000 UTC m=+1186.119476952" watchObservedRunningTime="2025-10-02 09:56:38.48607495 +0000 UTC m=+1186.133760017" Oct 02 09:56:38 crc kubenswrapper[4771]: I1002 09:56:38.625767 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" podStartSLOduration=6.466817056 podStartE2EDuration="37.625732012s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.598843079 +0000 UTC m=+1153.246528146" lastFinishedPulling="2025-10-02 09:56:36.757758035 +0000 UTC m=+1184.405443102" observedRunningTime="2025-10-02 09:56:38.623664399 +0000 UTC m=+1186.271349486" watchObservedRunningTime="2025-10-02 09:56:38.625732012 +0000 UTC m=+1186.273417079" Oct 02 09:56:38 crc kubenswrapper[4771]: E1002 09:56:38.681383 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.48:5001/openstack-k8s-operators/telemetry-operator:114f1d2f7a3b244c57adf7a27aa58cc483e58158\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" podUID="a8a3258c-981f-4f61-bb08-42b5065a9067" Oct 02 09:56:38 crc kubenswrapper[4771]: E1002 09:56:38.682049 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:21792a2317c0a55e40b2a02a7d5d4682b76538ed2a2e0633199aa395e60ecc72\\\"\"" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" podUID="6dc77080-a5ab-4db4-b150-2573589da332" Oct 02 09:56:38 crc kubenswrapper[4771]: E1002 09:56:38.682146 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" podUID="04a0b3f5-6e78-477a-8bd6-9250ba41eabf" Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.437032 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" event={"ID":"1f236f0a-744a-4f98-90e9-7c78ead31ddc","Type":"ContainerStarted","Data":"b597a22eca28e60a194d4d09bba5beddbae8c1f7a30d2f3c53c493a95dde4cb5"} Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.438405 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.441339 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6" event={"ID":"f43ba614-8bea-47db-be7d-10c7a4e9ddbb","Type":"ContainerStarted","Data":"0a09e17fb2b7558a66cfe344bbfceefec453722a0a797b6e46cfc2608e084b78"} Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.441545 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6" Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.444191 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5" event={"ID":"ac8e19b1-242e-4785-96ee-4c72ee975cce","Type":"ContainerStarted","Data":"e9b583fab20fd663a8c56d09561a9e3d49194c8ba06fbc00648ce2f93b5759c3"} Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.445049 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5" Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.449324 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" event={"ID":"7f64d01d-a2f3-4e96-9f81-13cafb80225e","Type":"ContainerStarted","Data":"9f1b75208c5f65790454f2341ab0a971d84d11f90c070bc20aefb18a97629f4c"} Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.450154 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.452454 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw" event={"ID":"5fcc6b1c-2163-454b-9172-16852b9f8966","Type":"ContainerStarted","Data":"fb1db4022451144ad140022dfebdc810e59190a621e80043e70cd3f60a851435"} Oct 02 09:56:39 crc kubenswrapper[4771]: E1002 09:56:39.455173 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.48:5001/openstack-k8s-operators/telemetry-operator:114f1d2f7a3b244c57adf7a27aa58cc483e58158\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" podUID="a8a3258c-981f-4f61-bb08-42b5065a9067" Oct 02 09:56:39 crc kubenswrapper[4771]: E1002 09:56:39.455201 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:21792a2317c0a55e40b2a02a7d5d4682b76538ed2a2e0633199aa395e60ecc72\\\"\"" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" podUID="6dc77080-a5ab-4db4-b150-2573589da332" Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.492547 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" podStartSLOduration=6.960004466 podStartE2EDuration="38.492514515s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.26998605 +0000 UTC m=+1152.917671117" lastFinishedPulling="2025-10-02 09:56:36.802496089 +0000 UTC m=+1184.450181166" observedRunningTime="2025-10-02 09:56:39.470286377 +0000 UTC m=+1187.117971464" watchObservedRunningTime="2025-10-02 09:56:39.492514515 +0000 UTC m=+1187.140199582" Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.497373 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw" podStartSLOduration=6.95862626 podStartE2EDuration="38.497355849s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:03.733990512 +0000 UTC m=+1151.381675580" lastFinishedPulling="2025-10-02 09:56:35.272720112 +0000 UTC m=+1182.920405169" observedRunningTime="2025-10-02 09:56:39.491176251 +0000 UTC m=+1187.138861328" watchObservedRunningTime="2025-10-02 09:56:39.497355849 +0000 UTC m=+1187.145040916" Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.549213 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6" podStartSLOduration=9.332720178 podStartE2EDuration="38.549177434s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.474557721 +0000 UTC m=+1153.122242788" lastFinishedPulling="2025-10-02 09:56:34.691014977 +0000 UTC m=+1182.338700044" observedRunningTime="2025-10-02 09:56:39.525657133 +0000 UTC m=+1187.173342210" watchObservedRunningTime="2025-10-02 09:56:39.549177434 +0000 UTC m=+1187.196862501" Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.580666 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5" podStartSLOduration=9.114953159 podStartE2EDuration="38.580630658s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.226090547 +0000 UTC m=+1152.873775624" lastFinishedPulling="2025-10-02 09:56:34.691768066 +0000 UTC m=+1182.339453123" observedRunningTime="2025-10-02 09:56:39.565777088 +0000 UTC m=+1187.213462155" watchObservedRunningTime="2025-10-02 09:56:39.580630658 +0000 UTC m=+1187.228315725" Oct 02 09:56:39 crc kubenswrapper[4771]: I1002 09:56:39.598671 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" podStartSLOduration=7.535338938 podStartE2EDuration="38.598643949s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.519520301 +0000 UTC m=+1153.167205378" lastFinishedPulling="2025-10-02 09:56:36.582825322 +0000 UTC m=+1184.230510389" observedRunningTime="2025-10-02 09:56:39.592126472 +0000 UTC m=+1187.239811539" watchObservedRunningTime="2025-10-02 09:56:39.598643949 +0000 UTC m=+1187.246329016" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.472079 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756" event={"ID":"9804f0f6-0273-4b09-830f-f082cf718803","Type":"ContainerStarted","Data":"cbd3a9943f0b96f779e24d556cd29822f45a20a6d155b412d2c4bea226e9b268"} Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.473230 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.479552 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw" event={"ID":"f0638dc6-d014-4a78-bcfa-eb4e5cea5caf","Type":"ContainerStarted","Data":"bc787de588c28ba53cbe9b4ac260ef141dee929af36261f341d541a7297cb4b5"} Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.479747 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.482310 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd" event={"ID":"4a399512-1d85-4f61-b8d4-800a57c33f4d","Type":"ContainerStarted","Data":"e0e92d03f8b6f5b817519b34bee3e4fc15b93561c1a5de8339385c670018b89f"} Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.482465 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.488343 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l" event={"ID":"7e625416-21a2-45e6-aec5-cf04c408e65e","Type":"ContainerStarted","Data":"30ad3cc064dcd2d27ab9290ddd6f844d5497903ed6860c784cc3d2ac72a33dd8"} Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.488511 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.492153 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz" event={"ID":"57aab528-153f-4c43-95d6-0c076d2071df","Type":"ContainerStarted","Data":"0b90c58750e216a2233430db0cafa8b26194edfd497511b611b4420b86191575"} Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.493016 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.498021 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" event={"ID":"e0df3275-79e0-49b8-93d5-585e2667a5c4","Type":"ContainerStarted","Data":"9bbcb27a21d46c8821d125de7a74107f0ffd09df8051e92a64cd658f560db25b"} Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.498211 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.501035 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj" event={"ID":"8e8d717b-41cc-4dd7-b9e3-4d06ac36cc54","Type":"ContainerStarted","Data":"06b878772df916355d6fccaadf649109f2a3aa04ce905669043234d8552f4e77"} Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.502094 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.503698 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756" podStartSLOduration=5.956467404 podStartE2EDuration="39.503679992s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.257221273 +0000 UTC m=+1152.904906340" lastFinishedPulling="2025-10-02 09:56:38.804433861 +0000 UTC m=+1186.452118928" observedRunningTime="2025-10-02 09:56:40.500644724 +0000 UTC m=+1188.148329791" watchObservedRunningTime="2025-10-02 09:56:40.503679992 +0000 UTC m=+1188.151365059" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.549045 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l" podStartSLOduration=9.982208766 podStartE2EDuration="39.549026221s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.124199702 +0000 UTC m=+1152.771884769" lastFinishedPulling="2025-10-02 09:56:34.691017157 +0000 UTC m=+1182.338702224" observedRunningTime="2025-10-02 09:56:40.544927286 +0000 UTC m=+1188.192612353" watchObservedRunningTime="2025-10-02 09:56:40.549026221 +0000 UTC m=+1188.196711278" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.566659 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd" podStartSLOduration=6.569056238 podStartE2EDuration="39.566616401s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.19686558 +0000 UTC m=+1152.844550637" lastFinishedPulling="2025-10-02 09:56:38.194425723 +0000 UTC m=+1185.842110800" observedRunningTime="2025-10-02 09:56:40.560643818 +0000 UTC m=+1188.208328885" watchObservedRunningTime="2025-10-02 09:56:40.566616401 +0000 UTC m=+1188.214301468" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.603152 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz" podStartSLOduration=4.913942957 podStartE2EDuration="39.603099234s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.124227933 +0000 UTC m=+1152.771913000" lastFinishedPulling="2025-10-02 09:56:39.81338421 +0000 UTC m=+1187.461069277" observedRunningTime="2025-10-02 09:56:40.598088886 +0000 UTC m=+1188.245773953" watchObservedRunningTime="2025-10-02 09:56:40.603099234 +0000 UTC m=+1188.250784301" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.630712 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw" podStartSLOduration=9.407847558 podStartE2EDuration="39.630681369s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:03.952695935 +0000 UTC m=+1151.600381012" lastFinishedPulling="2025-10-02 09:56:34.175529756 +0000 UTC m=+1181.823214823" observedRunningTime="2025-10-02 09:56:40.621684759 +0000 UTC m=+1188.269369836" watchObservedRunningTime="2025-10-02 09:56:40.630681369 +0000 UTC m=+1188.278366436" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.647819 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" podStartSLOduration=10.149641787 podStartE2EDuration="39.647787707s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.193005981 +0000 UTC m=+1152.840691048" lastFinishedPulling="2025-10-02 09:56:34.691151901 +0000 UTC m=+1182.338836968" observedRunningTime="2025-10-02 09:56:40.64636853 +0000 UTC m=+1188.294053597" watchObservedRunningTime="2025-10-02 09:56:40.647787707 +0000 UTC m=+1188.295472794" Oct 02 09:56:40 crc kubenswrapper[4771]: I1002 09:56:40.679318 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj" podStartSLOduration=4.826411446 podStartE2EDuration="39.679290652s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:03.953411553 +0000 UTC m=+1151.601096620" lastFinishedPulling="2025-10-02 09:56:38.806290759 +0000 UTC m=+1186.453975826" observedRunningTime="2025-10-02 09:56:40.672441627 +0000 UTC m=+1188.320126694" watchObservedRunningTime="2025-10-02 09:56:40.679290652 +0000 UTC m=+1188.326975719" Oct 02 09:56:41 crc kubenswrapper[4771]: I1002 09:56:41.513077 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk" event={"ID":"4442fde9-3b77-4013-a6c2-87cfeca2b299","Type":"ContainerStarted","Data":"e453df48cf7de5444a40c9ad121a8de52e6b6ab2cf7baeb20d01fa3b09f8c555"} Oct 02 09:56:41 crc kubenswrapper[4771]: I1002 09:56:41.513929 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk" Oct 02 09:56:41 crc kubenswrapper[4771]: I1002 09:56:41.516820 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5" event={"ID":"9b273823-7cb9-4c85-b5cc-54a52a5d04b8","Type":"ContainerStarted","Data":"a1cfb232db8dc34fd5520754c6414147608c13c22eaa79d2d62d0a53a024c54c"} Oct 02 09:56:41 crc kubenswrapper[4771]: I1002 09:56:41.517590 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5" Oct 02 09:56:41 crc kubenswrapper[4771]: I1002 09:56:41.520234 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p" event={"ID":"7c2eb266-7f51-45b9-8095-89d219bb59f0","Type":"ContainerStarted","Data":"ca69fbd081ebbab827160215f29f8ca220ac5d36769ddc6847c4d1276c4115e0"} Oct 02 09:56:41 crc kubenswrapper[4771]: I1002 09:56:41.520307 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p" Oct 02 09:56:41 crc kubenswrapper[4771]: I1002 09:56:41.527052 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h" event={"ID":"4da5dda4-f0da-453d-ba80-a2fc95035688","Type":"ContainerStarted","Data":"841d117bac326a67e384396250adc741f34be222820790faeade5a0a66582e19"} Oct 02 09:56:41 crc kubenswrapper[4771]: I1002 09:56:41.527935 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h" Oct 02 09:56:41 crc kubenswrapper[4771]: I1002 09:56:41.527977 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj" Oct 02 09:56:41 crc kubenswrapper[4771]: I1002 09:56:41.555230 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk" podStartSLOduration=5.526489549 podStartE2EDuration="40.55519446s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.198489401 +0000 UTC m=+1152.846174468" lastFinishedPulling="2025-10-02 09:56:40.227194312 +0000 UTC m=+1187.874879379" observedRunningTime="2025-10-02 09:56:41.542252129 +0000 UTC m=+1189.189937196" watchObservedRunningTime="2025-10-02 09:56:41.55519446 +0000 UTC m=+1189.202879527" Oct 02 09:56:41 crc kubenswrapper[4771]: I1002 09:56:41.614075 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p" podStartSLOduration=5.429600923 podStartE2EDuration="40.614044115s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.044067263 +0000 UTC m=+1152.691752330" lastFinishedPulling="2025-10-02 09:56:40.228510455 +0000 UTC m=+1187.876195522" observedRunningTime="2025-10-02 09:56:41.576015362 +0000 UTC m=+1189.223700429" watchObservedRunningTime="2025-10-02 09:56:41.614044115 +0000 UTC m=+1189.261729182" Oct 02 09:56:41 crc kubenswrapper[4771]: I1002 09:56:41.616979 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h" podStartSLOduration=4.304482393 podStartE2EDuration="40.61696713s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:03.917949537 +0000 UTC m=+1151.565634614" lastFinishedPulling="2025-10-02 09:56:40.230434284 +0000 UTC m=+1187.878119351" observedRunningTime="2025-10-02 09:56:41.609696744 +0000 UTC m=+1189.257381831" watchObservedRunningTime="2025-10-02 09:56:41.61696713 +0000 UTC m=+1189.264652197" Oct 02 09:56:41 crc kubenswrapper[4771]: I1002 09:56:41.635170 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5" podStartSLOduration=5.600244025 podStartE2EDuration="40.635103423s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.162190413 +0000 UTC m=+1152.809875480" lastFinishedPulling="2025-10-02 09:56:40.197049811 +0000 UTC m=+1187.844734878" observedRunningTime="2025-10-02 09:56:41.628157546 +0000 UTC m=+1189.275842633" watchObservedRunningTime="2025-10-02 09:56:41.635103423 +0000 UTC m=+1189.282788490" Oct 02 09:56:42 crc kubenswrapper[4771]: I1002 09:56:42.420808 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-gs7d5" Oct 02 09:56:42 crc kubenswrapper[4771]: I1002 09:56:42.667894 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-99mgr" Oct 02 09:56:42 crc kubenswrapper[4771]: I1002 09:56:42.708501 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-85777745bb-59m2p" Oct 02 09:56:42 crc kubenswrapper[4771]: I1002 09:56:42.827434 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-wmrf6" Oct 02 09:56:51 crc kubenswrapper[4771]: I1002 09:56:51.546680 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-smc8h" Oct 02 09:56:51 crc kubenswrapper[4771]: I1002 09:56:51.568773 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-dmwmw" Oct 02 09:56:51 crc kubenswrapper[4771]: I1002 09:56:51.604777 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-5xvqw" Oct 02 09:56:51 crc kubenswrapper[4771]: I1002 09:56:51.654119 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" event={"ID":"6dc77080-a5ab-4db4-b150-2573589da332","Type":"ContainerStarted","Data":"649968e07be26e0d938acfa4dd98bf30f032b3c6436fdc37a8b04fbb0dddc151"} Oct 02 09:56:51 crc kubenswrapper[4771]: I1002 09:56:51.655961 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" Oct 02 09:56:51 crc kubenswrapper[4771]: I1002 09:56:51.666032 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" event={"ID":"a8a3258c-981f-4f61-bb08-42b5065a9067","Type":"ContainerStarted","Data":"e0038d9b3640396b3ca616c6e2e5ea8174858dd95e5974e8c69826597f48204d"} Oct 02 09:56:51 crc kubenswrapper[4771]: I1002 09:56:51.666543 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" Oct 02 09:56:51 crc kubenswrapper[4771]: I1002 09:56:51.696784 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" podStartSLOduration=3.436343973 podStartE2EDuration="50.696763438s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:03.847131726 +0000 UTC m=+1151.494816803" lastFinishedPulling="2025-10-02 09:56:51.107551201 +0000 UTC m=+1198.755236268" observedRunningTime="2025-10-02 09:56:51.686141636 +0000 UTC m=+1199.333826713" watchObservedRunningTime="2025-10-02 09:56:51.696763438 +0000 UTC m=+1199.344448505" Oct 02 09:56:51 crc kubenswrapper[4771]: I1002 09:56:51.749998 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-hh756" Oct 02 09:56:51 crc kubenswrapper[4771]: I1002 09:56:51.793916 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" podStartSLOduration=5.223645995 podStartE2EDuration="50.793889411s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.197187278 +0000 UTC m=+1152.844872345" lastFinishedPulling="2025-10-02 09:56:50.767430694 +0000 UTC m=+1198.415115761" observedRunningTime="2025-10-02 09:56:51.734426411 +0000 UTC m=+1199.382111478" watchObservedRunningTime="2025-10-02 09:56:51.793889411 +0000 UTC m=+1199.441574478" Oct 02 09:56:51 crc kubenswrapper[4771]: I1002 09:56:51.832006 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-lp8xj" Oct 02 09:56:52 crc kubenswrapper[4771]: I1002 09:56:52.105540 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4nr9l" Oct 02 09:56:52 crc kubenswrapper[4771]: I1002 09:56:52.155413 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-tgx7p" Oct 02 09:56:52 crc kubenswrapper[4771]: I1002 09:56:52.157027 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jg6rv" Oct 02 09:56:52 crc kubenswrapper[4771]: I1002 09:56:52.229144 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zwjcd" Oct 02 09:56:52 crc kubenswrapper[4771]: I1002 09:56:52.517460 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-mrfv5" Oct 02 09:56:52 crc kubenswrapper[4771]: I1002 09:56:52.534787 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-4f4vk" Oct 02 09:56:52 crc kubenswrapper[4771]: I1002 09:56:52.607417 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-l76nz" Oct 02 09:56:52 crc kubenswrapper[4771]: I1002 09:56:52.679731 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" event={"ID":"04a0b3f5-6e78-477a-8bd6-9250ba41eabf","Type":"ContainerStarted","Data":"9503d541122366ad572627bf8455b4642557a99a6073a92281fee26299aacaa6"} Oct 02 09:56:52 crc kubenswrapper[4771]: I1002 09:56:52.680331 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" Oct 02 09:56:52 crc kubenswrapper[4771]: I1002 09:56:52.722110 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" podStartSLOduration=5.46042027 podStartE2EDuration="51.722084086s" podCreationTimestamp="2025-10-02 09:56:01 +0000 UTC" firstStartedPulling="2025-10-02 09:56:05.195971287 +0000 UTC m=+1152.843656364" lastFinishedPulling="2025-10-02 09:56:51.457635113 +0000 UTC m=+1199.105320180" observedRunningTime="2025-10-02 09:56:52.713678611 +0000 UTC m=+1200.361363678" watchObservedRunningTime="2025-10-02 09:56:52.722084086 +0000 UTC m=+1200.369769153" Oct 02 09:56:52 crc kubenswrapper[4771]: I1002 09:56:52.758718 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-hxpcp" Oct 02 09:56:52 crc kubenswrapper[4771]: I1002 09:56:52.950987 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-lpb8d" Oct 02 09:57:01 crc kubenswrapper[4771]: I1002 09:57:01.658245 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-pnqnd" Oct 02 09:57:02 crc kubenswrapper[4771]: I1002 09:57:02.130660 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-4dcd9" Oct 02 09:57:02 crc kubenswrapper[4771]: I1002 09:57:02.474864 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-d4856d54-q8sbj" Oct 02 09:57:18 crc kubenswrapper[4771]: I1002 09:57:18.785217 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-hbd2n"] Oct 02 09:57:18 crc kubenswrapper[4771]: I1002 09:57:18.787373 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-hbd2n" Oct 02 09:57:18 crc kubenswrapper[4771]: I1002 09:57:18.793116 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-ft2sg" Oct 02 09:57:18 crc kubenswrapper[4771]: I1002 09:57:18.793391 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 02 09:57:18 crc kubenswrapper[4771]: I1002 09:57:18.793400 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 02 09:57:18 crc kubenswrapper[4771]: I1002 09:57:18.799530 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 02 09:57:18 crc kubenswrapper[4771]: I1002 09:57:18.806693 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-hbd2n"] Oct 02 09:57:18 crc kubenswrapper[4771]: I1002 09:57:18.931994 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mczfg\" (UniqueName: \"kubernetes.io/projected/8a4bd248-50ba-48c2-8425-bd5e86b0e594-kube-api-access-mczfg\") pod \"dnsmasq-dns-675f4bcbfc-hbd2n\" (UID: \"8a4bd248-50ba-48c2-8425-bd5e86b0e594\") " pod="openstack/dnsmasq-dns-675f4bcbfc-hbd2n" Oct 02 09:57:18 crc kubenswrapper[4771]: I1002 09:57:18.932109 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a4bd248-50ba-48c2-8425-bd5e86b0e594-config\") pod \"dnsmasq-dns-675f4bcbfc-hbd2n\" (UID: \"8a4bd248-50ba-48c2-8425-bd5e86b0e594\") " pod="openstack/dnsmasq-dns-675f4bcbfc-hbd2n" Oct 02 09:57:18 crc kubenswrapper[4771]: I1002 09:57:18.940344 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2v2x2"] Oct 02 09:57:18 crc kubenswrapper[4771]: I1002 09:57:18.942896 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" Oct 02 09:57:18 crc kubenswrapper[4771]: I1002 09:57:18.945752 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 02 09:57:18 crc kubenswrapper[4771]: I1002 09:57:18.978711 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2v2x2"] Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.033789 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mczfg\" (UniqueName: \"kubernetes.io/projected/8a4bd248-50ba-48c2-8425-bd5e86b0e594-kube-api-access-mczfg\") pod \"dnsmasq-dns-675f4bcbfc-hbd2n\" (UID: \"8a4bd248-50ba-48c2-8425-bd5e86b0e594\") " pod="openstack/dnsmasq-dns-675f4bcbfc-hbd2n" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.033880 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a4bd248-50ba-48c2-8425-bd5e86b0e594-config\") pod \"dnsmasq-dns-675f4bcbfc-hbd2n\" (UID: \"8a4bd248-50ba-48c2-8425-bd5e86b0e594\") " pod="openstack/dnsmasq-dns-675f4bcbfc-hbd2n" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.033937 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/989a5396-f3cd-4231-a14d-fe7bac16db35-config\") pod \"dnsmasq-dns-78dd6ddcc-2v2x2\" (UID: \"989a5396-f3cd-4231-a14d-fe7bac16db35\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.034034 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/989a5396-f3cd-4231-a14d-fe7bac16db35-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2v2x2\" (UID: \"989a5396-f3cd-4231-a14d-fe7bac16db35\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.034057 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rspxq\" (UniqueName: \"kubernetes.io/projected/989a5396-f3cd-4231-a14d-fe7bac16db35-kube-api-access-rspxq\") pod \"dnsmasq-dns-78dd6ddcc-2v2x2\" (UID: \"989a5396-f3cd-4231-a14d-fe7bac16db35\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.035319 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a4bd248-50ba-48c2-8425-bd5e86b0e594-config\") pod \"dnsmasq-dns-675f4bcbfc-hbd2n\" (UID: \"8a4bd248-50ba-48c2-8425-bd5e86b0e594\") " pod="openstack/dnsmasq-dns-675f4bcbfc-hbd2n" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.066260 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mczfg\" (UniqueName: \"kubernetes.io/projected/8a4bd248-50ba-48c2-8425-bd5e86b0e594-kube-api-access-mczfg\") pod \"dnsmasq-dns-675f4bcbfc-hbd2n\" (UID: \"8a4bd248-50ba-48c2-8425-bd5e86b0e594\") " pod="openstack/dnsmasq-dns-675f4bcbfc-hbd2n" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.119753 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-hbd2n" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.136893 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/989a5396-f3cd-4231-a14d-fe7bac16db35-config\") pod \"dnsmasq-dns-78dd6ddcc-2v2x2\" (UID: \"989a5396-f3cd-4231-a14d-fe7bac16db35\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.137010 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/989a5396-f3cd-4231-a14d-fe7bac16db35-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2v2x2\" (UID: \"989a5396-f3cd-4231-a14d-fe7bac16db35\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.137047 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rspxq\" (UniqueName: \"kubernetes.io/projected/989a5396-f3cd-4231-a14d-fe7bac16db35-kube-api-access-rspxq\") pod \"dnsmasq-dns-78dd6ddcc-2v2x2\" (UID: \"989a5396-f3cd-4231-a14d-fe7bac16db35\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.138527 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/989a5396-f3cd-4231-a14d-fe7bac16db35-config\") pod \"dnsmasq-dns-78dd6ddcc-2v2x2\" (UID: \"989a5396-f3cd-4231-a14d-fe7bac16db35\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.139153 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/989a5396-f3cd-4231-a14d-fe7bac16db35-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2v2x2\" (UID: \"989a5396-f3cd-4231-a14d-fe7bac16db35\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.173055 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rspxq\" (UniqueName: \"kubernetes.io/projected/989a5396-f3cd-4231-a14d-fe7bac16db35-kube-api-access-rspxq\") pod \"dnsmasq-dns-78dd6ddcc-2v2x2\" (UID: \"989a5396-f3cd-4231-a14d-fe7bac16db35\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.262641 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.821869 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2v2x2"] Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.941582 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-hbd2n"] Oct 02 09:57:19 crc kubenswrapper[4771]: W1002 09:57:19.943947 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a4bd248_50ba_48c2_8425_bd5e86b0e594.slice/crio-5bbb43c5f1202e8eb22328cbe9556acc1e6ee143846b3ab03d3e6acfd87a2896 WatchSource:0}: Error finding container 5bbb43c5f1202e8eb22328cbe9556acc1e6ee143846b3ab03d3e6acfd87a2896: Status 404 returned error can't find the container with id 5bbb43c5f1202e8eb22328cbe9556acc1e6ee143846b3ab03d3e6acfd87a2896 Oct 02 09:57:19 crc kubenswrapper[4771]: I1002 09:57:19.944500 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" event={"ID":"989a5396-f3cd-4231-a14d-fe7bac16db35","Type":"ContainerStarted","Data":"9fa7f01913676fb4009f8f3bb9bf2c9917d4d8d15ba842a41de3f6df32495421"} Oct 02 09:57:20 crc kubenswrapper[4771]: I1002 09:57:20.958337 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-hbd2n" event={"ID":"8a4bd248-50ba-48c2-8425-bd5e86b0e594","Type":"ContainerStarted","Data":"5bbb43c5f1202e8eb22328cbe9556acc1e6ee143846b3ab03d3e6acfd87a2896"} Oct 02 09:57:21 crc kubenswrapper[4771]: I1002 09:57:21.849049 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-hbd2n"] Oct 02 09:57:21 crc kubenswrapper[4771]: I1002 09:57:21.874389 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-7hclw"] Oct 02 09:57:21 crc kubenswrapper[4771]: I1002 09:57:21.881209 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" Oct 02 09:57:21 crc kubenswrapper[4771]: I1002 09:57:21.885688 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-7hclw"] Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.013416 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4265ff2d-9844-4d4f-922a-efe637c14e8d-config\") pod \"dnsmasq-dns-5ccc8479f9-7hclw\" (UID: \"4265ff2d-9844-4d4f-922a-efe637c14e8d\") " pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.013458 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w474\" (UniqueName: \"kubernetes.io/projected/4265ff2d-9844-4d4f-922a-efe637c14e8d-kube-api-access-9w474\") pod \"dnsmasq-dns-5ccc8479f9-7hclw\" (UID: \"4265ff2d-9844-4d4f-922a-efe637c14e8d\") " pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.013601 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4265ff2d-9844-4d4f-922a-efe637c14e8d-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-7hclw\" (UID: \"4265ff2d-9844-4d4f-922a-efe637c14e8d\") " pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.116209 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4265ff2d-9844-4d4f-922a-efe637c14e8d-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-7hclw\" (UID: \"4265ff2d-9844-4d4f-922a-efe637c14e8d\") " pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.116349 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4265ff2d-9844-4d4f-922a-efe637c14e8d-config\") pod \"dnsmasq-dns-5ccc8479f9-7hclw\" (UID: \"4265ff2d-9844-4d4f-922a-efe637c14e8d\") " pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.116376 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w474\" (UniqueName: \"kubernetes.io/projected/4265ff2d-9844-4d4f-922a-efe637c14e8d-kube-api-access-9w474\") pod \"dnsmasq-dns-5ccc8479f9-7hclw\" (UID: \"4265ff2d-9844-4d4f-922a-efe637c14e8d\") " pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.117860 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4265ff2d-9844-4d4f-922a-efe637c14e8d-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-7hclw\" (UID: \"4265ff2d-9844-4d4f-922a-efe637c14e8d\") " pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.118002 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4265ff2d-9844-4d4f-922a-efe637c14e8d-config\") pod \"dnsmasq-dns-5ccc8479f9-7hclw\" (UID: \"4265ff2d-9844-4d4f-922a-efe637c14e8d\") " pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.147451 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w474\" (UniqueName: \"kubernetes.io/projected/4265ff2d-9844-4d4f-922a-efe637c14e8d-kube-api-access-9w474\") pod \"dnsmasq-dns-5ccc8479f9-7hclw\" (UID: \"4265ff2d-9844-4d4f-922a-efe637c14e8d\") " pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.225558 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.379424 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2v2x2"] Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.432720 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8h9mn"] Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.435054 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.465121 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8h9mn"] Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.529061 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0091b0aa-e43f-476c-8383-c3d279dc11d2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8h9mn\" (UID: \"0091b0aa-e43f-476c-8383-c3d279dc11d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.529302 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8jmq\" (UniqueName: \"kubernetes.io/projected/0091b0aa-e43f-476c-8383-c3d279dc11d2-kube-api-access-x8jmq\") pod \"dnsmasq-dns-57d769cc4f-8h9mn\" (UID: \"0091b0aa-e43f-476c-8383-c3d279dc11d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.529394 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0091b0aa-e43f-476c-8383-c3d279dc11d2-config\") pod \"dnsmasq-dns-57d769cc4f-8h9mn\" (UID: \"0091b0aa-e43f-476c-8383-c3d279dc11d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.630623 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0091b0aa-e43f-476c-8383-c3d279dc11d2-config\") pod \"dnsmasq-dns-57d769cc4f-8h9mn\" (UID: \"0091b0aa-e43f-476c-8383-c3d279dc11d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.631150 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0091b0aa-e43f-476c-8383-c3d279dc11d2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8h9mn\" (UID: \"0091b0aa-e43f-476c-8383-c3d279dc11d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.631298 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8jmq\" (UniqueName: \"kubernetes.io/projected/0091b0aa-e43f-476c-8383-c3d279dc11d2-kube-api-access-x8jmq\") pod \"dnsmasq-dns-57d769cc4f-8h9mn\" (UID: \"0091b0aa-e43f-476c-8383-c3d279dc11d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.632288 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0091b0aa-e43f-476c-8383-c3d279dc11d2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8h9mn\" (UID: \"0091b0aa-e43f-476c-8383-c3d279dc11d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.633073 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0091b0aa-e43f-476c-8383-c3d279dc11d2-config\") pod \"dnsmasq-dns-57d769cc4f-8h9mn\" (UID: \"0091b0aa-e43f-476c-8383-c3d279dc11d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.684582 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8jmq\" (UniqueName: \"kubernetes.io/projected/0091b0aa-e43f-476c-8383-c3d279dc11d2-kube-api-access-x8jmq\") pod \"dnsmasq-dns-57d769cc4f-8h9mn\" (UID: \"0091b0aa-e43f-476c-8383-c3d279dc11d2\") " pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:57:22 crc kubenswrapper[4771]: I1002 09:57:22.804550 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.055089 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.057231 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.073090 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.074609 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.074854 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.075999 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.076247 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-kk7ht" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.076309 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.076448 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.083152 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.143663 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-7hclw"] Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.149398 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7dad196-4ebc-43e0-b850-0820f2b072c9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.149453 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.149490 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.149665 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.149719 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.149813 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.149956 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.150020 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7dad196-4ebc-43e0-b850-0820f2b072c9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.150058 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.150214 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hz5b\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-kube-api-access-4hz5b\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.150571 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: W1002 09:57:23.155573 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4265ff2d_9844_4d4f_922a_efe637c14e8d.slice/crio-e7b6f882f88ce8ab6df3fd121a5f3ef2f1580273bdef98d635bba07406237587 WatchSource:0}: Error finding container e7b6f882f88ce8ab6df3fd121a5f3ef2f1580273bdef98d635bba07406237587: Status 404 returned error can't find the container with id e7b6f882f88ce8ab6df3fd121a5f3ef2f1580273bdef98d635bba07406237587 Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.255423 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.255492 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7dad196-4ebc-43e0-b850-0820f2b072c9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.255527 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.255598 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hz5b\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-kube-api-access-4hz5b\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.255803 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.256223 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.256438 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7dad196-4ebc-43e0-b850-0820f2b072c9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.256477 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.256510 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.256438 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.256545 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.256567 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.256622 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.257873 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.258000 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.258610 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.263250 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.263993 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7dad196-4ebc-43e0-b850-0820f2b072c9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.264377 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7dad196-4ebc-43e0-b850-0820f2b072c9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.275263 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.280024 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.287254 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hz5b\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-kube-api-access-4hz5b\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.324912 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.400333 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8h9mn"] Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.423317 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.566666 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.569404 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.577411 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.577557 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.577496 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.577969 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-v5mng" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.578014 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.578210 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.578362 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.601710 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.665847 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.665920 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.665944 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.666009 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.666032 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ebc075f5-e2ba-4789-a38e-4065fe661ada-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.666054 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.666091 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-config-data\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.666113 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnslr\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-kube-api-access-qnslr\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.666162 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.666197 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.666298 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ebc075f5-e2ba-4789-a38e-4065fe661ada-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.787983 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.788154 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.788187 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.788382 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.788411 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ebc075f5-e2ba-4789-a38e-4065fe661ada-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.788432 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.788500 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-config-data\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.788540 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnslr\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-kube-api-access-qnslr\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.788584 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.788644 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.788705 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ebc075f5-e2ba-4789-a38e-4065fe661ada-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.806646 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.807604 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-config-data\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.815841 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.817791 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.819915 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.820804 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.825549 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ebc075f5-e2ba-4789-a38e-4065fe661ada-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.842113 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnslr\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-kube-api-access-qnslr\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.866357 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.880517 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.919028 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ebc075f5-e2ba-4789-a38e-4065fe661ada-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.919347 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " pod="openstack/rabbitmq-server-0" Oct 02 09:57:23 crc kubenswrapper[4771]: I1002 09:57:23.920179 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 09:57:24 crc kubenswrapper[4771]: I1002 09:57:24.016397 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 09:57:24 crc kubenswrapper[4771]: I1002 09:57:24.026451 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" event={"ID":"4265ff2d-9844-4d4f-922a-efe637c14e8d","Type":"ContainerStarted","Data":"e7b6f882f88ce8ab6df3fd121a5f3ef2f1580273bdef98d635bba07406237587"} Oct 02 09:57:24 crc kubenswrapper[4771]: I1002 09:57:24.027856 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" event={"ID":"0091b0aa-e43f-476c-8383-c3d279dc11d2","Type":"ContainerStarted","Data":"d15415aa27347b5656dc264942fa8e8acfab550e93c74786133a67786249c044"} Oct 02 09:57:24 crc kubenswrapper[4771]: W1002 09:57:24.045005 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7dad196_4ebc_43e0_b850_0820f2b072c9.slice/crio-7bd65e96cfb2476a195caec04d490921c0aac5fba96d01ba8a91363fb21bcb5a WatchSource:0}: Error finding container 7bd65e96cfb2476a195caec04d490921c0aac5fba96d01ba8a91363fb21bcb5a: Status 404 returned error can't find the container with id 7bd65e96cfb2476a195caec04d490921c0aac5fba96d01ba8a91363fb21bcb5a Oct 02 09:57:24 crc kubenswrapper[4771]: I1002 09:57:24.644182 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 09:57:24 crc kubenswrapper[4771]: I1002 09:57:24.934009 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 02 09:57:24 crc kubenswrapper[4771]: I1002 09:57:24.937758 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 09:57:24 crc kubenswrapper[4771]: I1002 09:57:24.948704 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-sr5x5" Oct 02 09:57:24 crc kubenswrapper[4771]: I1002 09:57:24.948894 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 02 09:57:24 crc kubenswrapper[4771]: I1002 09:57:24.950184 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 02 09:57:24 crc kubenswrapper[4771]: I1002 09:57:24.950267 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 02 09:57:24 crc kubenswrapper[4771]: I1002 09:57:24.950184 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 02 09:57:24 crc kubenswrapper[4771]: I1002 09:57:24.970800 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.027640 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.057879 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/774a9026-9330-4c75-a2ca-33d5bccaf388-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.058018 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/774a9026-9330-4c75-a2ca-33d5bccaf388-config-data-default\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.058057 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.058099 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/774a9026-9330-4c75-a2ca-33d5bccaf388-secrets\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.058198 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/774a9026-9330-4c75-a2ca-33d5bccaf388-config-data-generated\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.058246 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/774a9026-9330-4c75-a2ca-33d5bccaf388-kolla-config\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.058269 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/774a9026-9330-4c75-a2ca-33d5bccaf388-operator-scripts\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.058310 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6kqg\" (UniqueName: \"kubernetes.io/projected/774a9026-9330-4c75-a2ca-33d5bccaf388-kube-api-access-f6kqg\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.058427 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/774a9026-9330-4c75-a2ca-33d5bccaf388-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.136010 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ebc075f5-e2ba-4789-a38e-4065fe661ada","Type":"ContainerStarted","Data":"198faa5123c5bbbb5493a576910da7893b55dcc83131eb1be938faa0e81b0ab2"} Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.152516 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7dad196-4ebc-43e0-b850-0820f2b072c9","Type":"ContainerStarted","Data":"7bd65e96cfb2476a195caec04d490921c0aac5fba96d01ba8a91363fb21bcb5a"} Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.161580 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/774a9026-9330-4c75-a2ca-33d5bccaf388-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.161697 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/774a9026-9330-4c75-a2ca-33d5bccaf388-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.161744 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/774a9026-9330-4c75-a2ca-33d5bccaf388-config-data-default\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.161775 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.161798 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/774a9026-9330-4c75-a2ca-33d5bccaf388-secrets\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.161876 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/774a9026-9330-4c75-a2ca-33d5bccaf388-config-data-generated\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.161906 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/774a9026-9330-4c75-a2ca-33d5bccaf388-kolla-config\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.161926 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/774a9026-9330-4c75-a2ca-33d5bccaf388-operator-scripts\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.161950 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6kqg\" (UniqueName: \"kubernetes.io/projected/774a9026-9330-4c75-a2ca-33d5bccaf388-kube-api-access-f6kqg\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.175638 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/774a9026-9330-4c75-a2ca-33d5bccaf388-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.175981 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/774a9026-9330-4c75-a2ca-33d5bccaf388-config-data-generated\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.176671 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/774a9026-9330-4c75-a2ca-33d5bccaf388-kolla-config\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.176927 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/774a9026-9330-4c75-a2ca-33d5bccaf388-config-data-default\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.177728 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/774a9026-9330-4c75-a2ca-33d5bccaf388-secrets\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.177938 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/774a9026-9330-4c75-a2ca-33d5bccaf388-operator-scripts\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.178104 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.182859 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/774a9026-9330-4c75-a2ca-33d5bccaf388-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.230398 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6kqg\" (UniqueName: \"kubernetes.io/projected/774a9026-9330-4c75-a2ca-33d5bccaf388-kube-api-access-f6kqg\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.274843 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"774a9026-9330-4c75-a2ca-33d5bccaf388\") " pod="openstack/openstack-galera-0" Oct 02 09:57:25 crc kubenswrapper[4771]: I1002 09:57:25.309590 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.199960 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.212799 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.217073 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.217102 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-pvrlf" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.217844 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.217987 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.224860 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.371574 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.373255 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.383282 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.383460 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-ts828" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.383626 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.400233 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/36a32369-c748-4356-be1b-5008d4f9bbe3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.400340 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/36a32369-c748-4356-be1b-5008d4f9bbe3-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.400385 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36a32369-c748-4356-be1b-5008d4f9bbe3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.400468 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36a32369-c748-4356-be1b-5008d4f9bbe3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.400507 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/36a32369-c748-4356-be1b-5008d4f9bbe3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.400530 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.400557 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/36a32369-c748-4356-be1b-5008d4f9bbe3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.400581 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/36a32369-c748-4356-be1b-5008d4f9bbe3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.400650 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bkks\" (UniqueName: \"kubernetes.io/projected/36a32369-c748-4356-be1b-5008d4f9bbe3-kube-api-access-6bkks\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.419158 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.502581 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bkks\" (UniqueName: \"kubernetes.io/projected/36a32369-c748-4356-be1b-5008d4f9bbe3-kube-api-access-6bkks\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.502649 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvb24\" (UniqueName: \"kubernetes.io/projected/785673c9-400c-474f-915d-30e8da38b311-kube-api-access-tvb24\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.502693 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/36a32369-c748-4356-be1b-5008d4f9bbe3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.502750 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/785673c9-400c-474f-915d-30e8da38b311-config-data\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.502774 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/36a32369-c748-4356-be1b-5008d4f9bbe3-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.502805 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36a32369-c748-4356-be1b-5008d4f9bbe3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.502900 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36a32369-c748-4356-be1b-5008d4f9bbe3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.502929 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/36a32369-c748-4356-be1b-5008d4f9bbe3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.502950 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.502971 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/36a32369-c748-4356-be1b-5008d4f9bbe3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.502988 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/36a32369-c748-4356-be1b-5008d4f9bbe3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.503021 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/785673c9-400c-474f-915d-30e8da38b311-memcached-tls-certs\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.503049 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/785673c9-400c-474f-915d-30e8da38b311-kolla-config\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.503505 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/785673c9-400c-474f-915d-30e8da38b311-combined-ca-bundle\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.512018 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/36a32369-c748-4356-be1b-5008d4f9bbe3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.512749 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.533228 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/36a32369-c748-4356-be1b-5008d4f9bbe3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.533347 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/36a32369-c748-4356-be1b-5008d4f9bbe3-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.533924 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36a32369-c748-4356-be1b-5008d4f9bbe3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.541883 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36a32369-c748-4356-be1b-5008d4f9bbe3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.543008 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/36a32369-c748-4356-be1b-5008d4f9bbe3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.556028 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.591052 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/36a32369-c748-4356-be1b-5008d4f9bbe3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.593443 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.594819 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bkks\" (UniqueName: \"kubernetes.io/projected/36a32369-c748-4356-be1b-5008d4f9bbe3-kube-api-access-6bkks\") pod \"openstack-cell1-galera-0\" (UID: \"36a32369-c748-4356-be1b-5008d4f9bbe3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.605576 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/785673c9-400c-474f-915d-30e8da38b311-memcached-tls-certs\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.605751 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/785673c9-400c-474f-915d-30e8da38b311-kolla-config\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.605842 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/785673c9-400c-474f-915d-30e8da38b311-combined-ca-bundle\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.605927 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvb24\" (UniqueName: \"kubernetes.io/projected/785673c9-400c-474f-915d-30e8da38b311-kube-api-access-tvb24\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.606050 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/785673c9-400c-474f-915d-30e8da38b311-config-data\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.607043 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/785673c9-400c-474f-915d-30e8da38b311-config-data\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.607437 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/785673c9-400c-474f-915d-30e8da38b311-kolla-config\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.614417 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/785673c9-400c-474f-915d-30e8da38b311-memcached-tls-certs\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.620585 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/785673c9-400c-474f-915d-30e8da38b311-combined-ca-bundle\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.639996 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvb24\" (UniqueName: \"kubernetes.io/projected/785673c9-400c-474f-915d-30e8da38b311-kube-api-access-tvb24\") pod \"memcached-0\" (UID: \"785673c9-400c-474f-915d-30e8da38b311\") " pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.752579 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 09:57:26 crc kubenswrapper[4771]: I1002 09:57:26.857709 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 09:57:27 crc kubenswrapper[4771]: I1002 09:57:27.269961 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"774a9026-9330-4c75-a2ca-33d5bccaf388","Type":"ContainerStarted","Data":"bca431579cd15a06ba563773a6da6553851e74fede185684a5322b391b8383ba"} Oct 02 09:57:27 crc kubenswrapper[4771]: I1002 09:57:27.755056 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 09:57:27 crc kubenswrapper[4771]: I1002 09:57:27.989824 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 09:57:28 crc kubenswrapper[4771]: W1002 09:57:28.023264 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36a32369_c748_4356_be1b_5008d4f9bbe3.slice/crio-b3592009d71632035de0a45217a336ab23706475f1350dbee57985ba865788dc WatchSource:0}: Error finding container b3592009d71632035de0a45217a336ab23706475f1350dbee57985ba865788dc: Status 404 returned error can't find the container with id b3592009d71632035de0a45217a336ab23706475f1350dbee57985ba865788dc Oct 02 09:57:28 crc kubenswrapper[4771]: I1002 09:57:28.317784 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 09:57:28 crc kubenswrapper[4771]: I1002 09:57:28.321294 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 09:57:28 crc kubenswrapper[4771]: I1002 09:57:28.337631 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-zf45h" Oct 02 09:57:28 crc kubenswrapper[4771]: I1002 09:57:28.369053 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 09:57:28 crc kubenswrapper[4771]: I1002 09:57:28.481370 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kcvh\" (UniqueName: \"kubernetes.io/projected/d8f54e64-697d-41b5-83fd-295ce5143d04-kube-api-access-8kcvh\") pod \"kube-state-metrics-0\" (UID: \"d8f54e64-697d-41b5-83fd-295ce5143d04\") " pod="openstack/kube-state-metrics-0" Oct 02 09:57:28 crc kubenswrapper[4771]: I1002 09:57:28.488387 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"785673c9-400c-474f-915d-30e8da38b311","Type":"ContainerStarted","Data":"85fceb538d96440bf4c7ff3e68eefee0fc6b204fefb2800f7de201113e6a3e43"} Oct 02 09:57:28 crc kubenswrapper[4771]: I1002 09:57:28.513206 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"36a32369-c748-4356-be1b-5008d4f9bbe3","Type":"ContainerStarted","Data":"b3592009d71632035de0a45217a336ab23706475f1350dbee57985ba865788dc"} Oct 02 09:57:28 crc kubenswrapper[4771]: I1002 09:57:28.588430 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kcvh\" (UniqueName: \"kubernetes.io/projected/d8f54e64-697d-41b5-83fd-295ce5143d04-kube-api-access-8kcvh\") pod \"kube-state-metrics-0\" (UID: \"d8f54e64-697d-41b5-83fd-295ce5143d04\") " pod="openstack/kube-state-metrics-0" Oct 02 09:57:28 crc kubenswrapper[4771]: I1002 09:57:28.655259 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kcvh\" (UniqueName: \"kubernetes.io/projected/d8f54e64-697d-41b5-83fd-295ce5143d04-kube-api-access-8kcvh\") pod \"kube-state-metrics-0\" (UID: \"d8f54e64-697d-41b5-83fd-295ce5143d04\") " pod="openstack/kube-state-metrics-0" Oct 02 09:57:28 crc kubenswrapper[4771]: I1002 09:57:28.678912 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.190028 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq"] Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.192209 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.195790 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.195786 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-dh8dj" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.214174 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq"] Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.326093 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7636c0e3-830f-4f22-88ca-52a4e1358e08-serving-cert\") pod \"observability-ui-dashboards-6584dc9448-9vbgq\" (UID: \"7636c0e3-830f-4f22-88ca-52a4e1358e08\") " pod="openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.326313 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjlgq\" (UniqueName: \"kubernetes.io/projected/7636c0e3-830f-4f22-88ca-52a4e1358e08-kube-api-access-pjlgq\") pod \"observability-ui-dashboards-6584dc9448-9vbgq\" (UID: \"7636c0e3-830f-4f22-88ca-52a4e1358e08\") " pod="openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.428651 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7636c0e3-830f-4f22-88ca-52a4e1358e08-serving-cert\") pod \"observability-ui-dashboards-6584dc9448-9vbgq\" (UID: \"7636c0e3-830f-4f22-88ca-52a4e1358e08\") " pod="openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq" Oct 02 09:57:29 crc kubenswrapper[4771]: E1002 09:57:29.432698 4771 secret.go:188] Couldn't get secret openshift-operators/observability-ui-dashboards: secret "observability-ui-dashboards" not found Oct 02 09:57:29 crc kubenswrapper[4771]: E1002 09:57:29.432848 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7636c0e3-830f-4f22-88ca-52a4e1358e08-serving-cert podName:7636c0e3-830f-4f22-88ca-52a4e1358e08 nodeName:}" failed. No retries permitted until 2025-10-02 09:57:29.932816519 +0000 UTC m=+1237.580501596 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/7636c0e3-830f-4f22-88ca-52a4e1358e08-serving-cert") pod "observability-ui-dashboards-6584dc9448-9vbgq" (UID: "7636c0e3-830f-4f22-88ca-52a4e1358e08") : secret "observability-ui-dashboards" not found Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.433665 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjlgq\" (UniqueName: \"kubernetes.io/projected/7636c0e3-830f-4f22-88ca-52a4e1358e08-kube-api-access-pjlgq\") pod \"observability-ui-dashboards-6584dc9448-9vbgq\" (UID: \"7636c0e3-830f-4f22-88ca-52a4e1358e08\") " pod="openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.487365 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjlgq\" (UniqueName: \"kubernetes.io/projected/7636c0e3-830f-4f22-88ca-52a4e1358e08-kube-api-access-pjlgq\") pod \"observability-ui-dashboards-6584dc9448-9vbgq\" (UID: \"7636c0e3-830f-4f22-88ca-52a4e1358e08\") " pod="openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.656186 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.668641 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.674448 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.674926 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.675070 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.675237 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-656s8" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.685670 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.769403 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.819961 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.923078 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.923168 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.923221 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f33acc08-4fd4-473f-9263-ec4841f48809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f33acc08-4fd4-473f-9263-ec4841f48809\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.923247 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.923392 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.923447 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.923494 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lknt7\" (UniqueName: \"kubernetes.io/projected/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-kube-api-access-lknt7\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.923552 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-config\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.974968 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5556b8d8b5-fj9fv"] Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.977028 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:29 crc kubenswrapper[4771]: I1002 09:57:29.994434 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5556b8d8b5-fj9fv"] Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.024831 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.025089 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.025271 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lknt7\" (UniqueName: \"kubernetes.io/projected/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-kube-api-access-lknt7\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.025397 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-config\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.025476 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.025565 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.025658 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f33acc08-4fd4-473f-9263-ec4841f48809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f33acc08-4fd4-473f-9263-ec4841f48809\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.025786 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.025893 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7636c0e3-830f-4f22-88ca-52a4e1358e08-serving-cert\") pod \"observability-ui-dashboards-6584dc9448-9vbgq\" (UID: \"7636c0e3-830f-4f22-88ca-52a4e1358e08\") " pod="openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq" Oct 02 09:57:30 crc kubenswrapper[4771]: E1002 09:57:30.026095 4771 secret.go:188] Couldn't get secret openshift-operators/observability-ui-dashboards: secret "observability-ui-dashboards" not found Oct 02 09:57:30 crc kubenswrapper[4771]: E1002 09:57:30.026251 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7636c0e3-830f-4f22-88ca-52a4e1358e08-serving-cert podName:7636c0e3-830f-4f22-88ca-52a4e1358e08 nodeName:}" failed. No retries permitted until 2025-10-02 09:57:31.026235474 +0000 UTC m=+1238.673920541 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/7636c0e3-830f-4f22-88ca-52a4e1358e08-serving-cert") pod "observability-ui-dashboards-6584dc9448-9vbgq" (UID: "7636c0e3-830f-4f22-88ca-52a4e1358e08") : secret "observability-ui-dashboards" not found Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.043400 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.026166 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.050605 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-config\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.050607 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.062716 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.065085 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lknt7\" (UniqueName: \"kubernetes.io/projected/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-kube-api-access-lknt7\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.072086 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.105013 4771 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.105058 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f33acc08-4fd4-473f-9263-ec4841f48809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f33acc08-4fd4-473f-9263-ec4841f48809\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5158297136c2a151404177ed2b564800e7d4895241ff310a5747c375183a3a49/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.128849 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfzv2\" (UniqueName: \"kubernetes.io/projected/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-kube-api-access-hfzv2\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.128884 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-oauth-serving-cert\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.128901 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-console-config\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.128926 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-trusted-ca-bundle\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.128947 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-service-ca\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.128992 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-console-serving-cert\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.129210 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-console-oauth-config\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.189224 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f33acc08-4fd4-473f-9263-ec4841f48809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f33acc08-4fd4-473f-9263-ec4841f48809\") pod \"prometheus-metric-storage-0\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.235175 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfzv2\" (UniqueName: \"kubernetes.io/projected/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-kube-api-access-hfzv2\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.235253 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-oauth-serving-cert\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.235286 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-console-config\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.236842 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-oauth-serving-cert\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.240360 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-trusted-ca-bundle\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.240514 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-service-ca\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.240736 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-console-serving-cert\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.241202 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-console-oauth-config\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.241879 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-trusted-ca-bundle\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.242858 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-service-ca\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.244831 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-console-config\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.245052 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-console-serving-cert\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.250875 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-console-oauth-config\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.271811 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfzv2\" (UniqueName: \"kubernetes.io/projected/96c8168c-bf7e-4307-b30c-f4e5dfbebdfe-kube-api-access-hfzv2\") pod \"console-5556b8d8b5-fj9fv\" (UID: \"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe\") " pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.320063 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.346760 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.414976 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 09:57:30 crc kubenswrapper[4771]: I1002 09:57:30.701951 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d8f54e64-697d-41b5-83fd-295ce5143d04","Type":"ContainerStarted","Data":"a23f78a1259dbb38e2e04485fa46cc26df896546b85ced9b3773c58baa2d17e3"} Oct 02 09:57:31 crc kubenswrapper[4771]: I1002 09:57:31.105264 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7636c0e3-830f-4f22-88ca-52a4e1358e08-serving-cert\") pod \"observability-ui-dashboards-6584dc9448-9vbgq\" (UID: \"7636c0e3-830f-4f22-88ca-52a4e1358e08\") " pod="openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq" Oct 02 09:57:31 crc kubenswrapper[4771]: I1002 09:57:31.130679 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7636c0e3-830f-4f22-88ca-52a4e1358e08-serving-cert\") pod \"observability-ui-dashboards-6584dc9448-9vbgq\" (UID: \"7636c0e3-830f-4f22-88ca-52a4e1358e08\") " pod="openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq" Oct 02 09:57:31 crc kubenswrapper[4771]: I1002 09:57:31.329631 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq" Oct 02 09:57:31 crc kubenswrapper[4771]: I1002 09:57:31.385644 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 09:57:31 crc kubenswrapper[4771]: I1002 09:57:31.555920 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5556b8d8b5-fj9fv"] Oct 02 09:57:31 crc kubenswrapper[4771]: I1002 09:57:31.824559 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5556b8d8b5-fj9fv" event={"ID":"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe","Type":"ContainerStarted","Data":"7aad8fedbf550e98cc6476304f6d68ca968ee5509feeaef97a696df5bfe0cead"} Oct 02 09:57:31 crc kubenswrapper[4771]: I1002 09:57:31.838729 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"95cd3d5a-1000-4212-b4d9-5d0da549c7f3","Type":"ContainerStarted","Data":"5760d3a33afe8aa8fab457b68102811193f6ca9901b26714b020c370de834fa6"} Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.053781 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq"] Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.584880 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5jf58"] Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.586617 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.589348 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-65smm" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.590145 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.590364 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.631558 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5jf58"] Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.673826 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3388e290-ed26-4869-b6ce-c93670e7d9cf-combined-ca-bundle\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.673935 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3388e290-ed26-4869-b6ce-c93670e7d9cf-var-log-ovn\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.676697 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bclpf\" (UniqueName: \"kubernetes.io/projected/3388e290-ed26-4869-b6ce-c93670e7d9cf-kube-api-access-bclpf\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.676758 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3388e290-ed26-4869-b6ce-c93670e7d9cf-var-run\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.676844 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3388e290-ed26-4869-b6ce-c93670e7d9cf-ovn-controller-tls-certs\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.676873 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3388e290-ed26-4869-b6ce-c93670e7d9cf-var-run-ovn\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.676950 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3388e290-ed26-4869-b6ce-c93670e7d9cf-scripts\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.689714 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-9qhq9"] Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.692843 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.746614 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-9qhq9"] Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.787677 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3388e290-ed26-4869-b6ce-c93670e7d9cf-combined-ca-bundle\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.787853 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3388e290-ed26-4869-b6ce-c93670e7d9cf-var-log-ovn\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.788093 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bclpf\" (UniqueName: \"kubernetes.io/projected/3388e290-ed26-4869-b6ce-c93670e7d9cf-kube-api-access-bclpf\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.788200 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3388e290-ed26-4869-b6ce-c93670e7d9cf-var-run\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.788300 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3388e290-ed26-4869-b6ce-c93670e7d9cf-ovn-controller-tls-certs\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.788346 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3388e290-ed26-4869-b6ce-c93670e7d9cf-var-run-ovn\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.788463 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3388e290-ed26-4869-b6ce-c93670e7d9cf-scripts\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.793472 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3388e290-ed26-4869-b6ce-c93670e7d9cf-var-log-ovn\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.796573 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3388e290-ed26-4869-b6ce-c93670e7d9cf-var-run\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.799180 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3388e290-ed26-4869-b6ce-c93670e7d9cf-scripts\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.821676 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3388e290-ed26-4869-b6ce-c93670e7d9cf-ovn-controller-tls-certs\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.829747 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3388e290-ed26-4869-b6ce-c93670e7d9cf-var-run-ovn\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.831810 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bclpf\" (UniqueName: \"kubernetes.io/projected/3388e290-ed26-4869-b6ce-c93670e7d9cf-kube-api-access-bclpf\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.833015 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3388e290-ed26-4869-b6ce-c93670e7d9cf-combined-ca-bundle\") pod \"ovn-controller-5jf58\" (UID: \"3388e290-ed26-4869-b6ce-c93670e7d9cf\") " pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.891599 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f3a8d6f9-c114-49ea-af29-8399e6a598cc-etc-ovs\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.892189 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3a8d6f9-c114-49ea-af29-8399e6a598cc-scripts\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.893547 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f3a8d6f9-c114-49ea-af29-8399e6a598cc-var-log\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.893608 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f3a8d6f9-c114-49ea-af29-8399e6a598cc-var-run\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.893863 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hndz\" (UniqueName: \"kubernetes.io/projected/f3a8d6f9-c114-49ea-af29-8399e6a598cc-kube-api-access-8hndz\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.893923 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f3a8d6f9-c114-49ea-af29-8399e6a598cc-var-lib\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.956994 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5jf58" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.996445 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f3a8d6f9-c114-49ea-af29-8399e6a598cc-etc-ovs\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.996531 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3a8d6f9-c114-49ea-af29-8399e6a598cc-scripts\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.996576 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f3a8d6f9-c114-49ea-af29-8399e6a598cc-var-log\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.996604 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f3a8d6f9-c114-49ea-af29-8399e6a598cc-var-run\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.996693 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hndz\" (UniqueName: \"kubernetes.io/projected/f3a8d6f9-c114-49ea-af29-8399e6a598cc-kube-api-access-8hndz\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.996719 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f3a8d6f9-c114-49ea-af29-8399e6a598cc-var-lib\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.997312 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f3a8d6f9-c114-49ea-af29-8399e6a598cc-var-lib\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.997464 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f3a8d6f9-c114-49ea-af29-8399e6a598cc-etc-ovs\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.998161 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f3a8d6f9-c114-49ea-af29-8399e6a598cc-var-run\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:32 crc kubenswrapper[4771]: I1002 09:57:32.998465 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f3a8d6f9-c114-49ea-af29-8399e6a598cc-var-log\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.001395 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3a8d6f9-c114-49ea-af29-8399e6a598cc-scripts\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.043456 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hndz\" (UniqueName: \"kubernetes.io/projected/f3a8d6f9-c114-49ea-af29-8399e6a598cc-kube-api-access-8hndz\") pod \"ovn-controller-ovs-9qhq9\" (UID: \"f3a8d6f9-c114-49ea-af29-8399e6a598cc\") " pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.274371 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.285782 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.288956 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.289214 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.289260 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.289407 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-2j4hs" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.290069 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.308749 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.334200 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.420553 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.420611 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfd7c23d-a847-44c8-8e36-01ed1520831d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.420664 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfd7c23d-a847-44c8-8e36-01ed1520831d-config\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.420712 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfd7c23d-a847-44c8-8e36-01ed1520831d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.420773 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cfd7c23d-a847-44c8-8e36-01ed1520831d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.420792 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lws6d\" (UniqueName: \"kubernetes.io/projected/cfd7c23d-a847-44c8-8e36-01ed1520831d-kube-api-access-lws6d\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.420825 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cfd7c23d-a847-44c8-8e36-01ed1520831d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.420875 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd7c23d-a847-44c8-8e36-01ed1520831d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.523376 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd7c23d-a847-44c8-8e36-01ed1520831d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.523496 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.523532 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfd7c23d-a847-44c8-8e36-01ed1520831d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.523612 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfd7c23d-a847-44c8-8e36-01ed1520831d-config\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.523699 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfd7c23d-a847-44c8-8e36-01ed1520831d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.523812 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cfd7c23d-a847-44c8-8e36-01ed1520831d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.523837 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lws6d\" (UniqueName: \"kubernetes.io/projected/cfd7c23d-a847-44c8-8e36-01ed1520831d-kube-api-access-lws6d\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.523894 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cfd7c23d-a847-44c8-8e36-01ed1520831d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.525206 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.525325 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cfd7c23d-a847-44c8-8e36-01ed1520831d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.525691 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cfd7c23d-a847-44c8-8e36-01ed1520831d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.537077 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfd7c23d-a847-44c8-8e36-01ed1520831d-config\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.547643 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd7c23d-a847-44c8-8e36-01ed1520831d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.550891 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfd7c23d-a847-44c8-8e36-01ed1520831d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.561375 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfd7c23d-a847-44c8-8e36-01ed1520831d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.575774 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lws6d\" (UniqueName: \"kubernetes.io/projected/cfd7c23d-a847-44c8-8e36-01ed1520831d-kube-api-access-lws6d\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.622514 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"cfd7c23d-a847-44c8-8e36-01ed1520831d\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:33 crc kubenswrapper[4771]: I1002 09:57:33.632708 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.318429 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.324571 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.331464 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.331675 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.331807 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-6ncst" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.332009 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.354932 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.411879 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.411926 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/795a79a7-e29c-496f-87c0-0bc8a1ef1570-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.412018 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/795a79a7-e29c-496f-87c0-0bc8a1ef1570-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.412072 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssz2m\" (UniqueName: \"kubernetes.io/projected/795a79a7-e29c-496f-87c0-0bc8a1ef1570-kube-api-access-ssz2m\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.412098 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/795a79a7-e29c-496f-87c0-0bc8a1ef1570-config\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.412114 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795a79a7-e29c-496f-87c0-0bc8a1ef1570-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.412166 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/795a79a7-e29c-496f-87c0-0bc8a1ef1570-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.412194 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/795a79a7-e29c-496f-87c0-0bc8a1ef1570-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.514451 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.514509 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/795a79a7-e29c-496f-87c0-0bc8a1ef1570-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.514614 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/795a79a7-e29c-496f-87c0-0bc8a1ef1570-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.514669 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssz2m\" (UniqueName: \"kubernetes.io/projected/795a79a7-e29c-496f-87c0-0bc8a1ef1570-kube-api-access-ssz2m\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.514694 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/795a79a7-e29c-496f-87c0-0bc8a1ef1570-config\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.514713 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795a79a7-e29c-496f-87c0-0bc8a1ef1570-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.514754 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/795a79a7-e29c-496f-87c0-0bc8a1ef1570-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.514788 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/795a79a7-e29c-496f-87c0-0bc8a1ef1570-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.515231 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.515785 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/795a79a7-e29c-496f-87c0-0bc8a1ef1570-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.515785 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/795a79a7-e29c-496f-87c0-0bc8a1ef1570-config\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.516062 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/795a79a7-e29c-496f-87c0-0bc8a1ef1570-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.543738 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/795a79a7-e29c-496f-87c0-0bc8a1ef1570-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.546663 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795a79a7-e29c-496f-87c0-0bc8a1ef1570-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.554566 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssz2m\" (UniqueName: \"kubernetes.io/projected/795a79a7-e29c-496f-87c0-0bc8a1ef1570-kube-api-access-ssz2m\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.560691 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.588680 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/795a79a7-e29c-496f-87c0-0bc8a1ef1570-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"795a79a7-e29c-496f-87c0-0bc8a1ef1570\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:35 crc kubenswrapper[4771]: I1002 09:57:35.676630 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 09:57:41 crc kubenswrapper[4771]: W1002 09:57:41.946388 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7636c0e3_830f_4f22_88ca_52a4e1358e08.slice/crio-b39553c7f312d617b9eb2337ed89cd7f4e747eb36908fd7d295c8c24d255ba86 WatchSource:0}: Error finding container b39553c7f312d617b9eb2337ed89cd7f4e747eb36908fd7d295c8c24d255ba86: Status 404 returned error can't find the container with id b39553c7f312d617b9eb2337ed89cd7f4e747eb36908fd7d295c8c24d255ba86 Oct 02 09:57:42 crc kubenswrapper[4771]: I1002 09:57:42.043812 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq" event={"ID":"7636c0e3-830f-4f22-88ca-52a4e1358e08","Type":"ContainerStarted","Data":"b39553c7f312d617b9eb2337ed89cd7f4e747eb36908fd7d295c8c24d255ba86"} Oct 02 09:57:43 crc kubenswrapper[4771]: I1002 09:57:43.056755 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5556b8d8b5-fj9fv" event={"ID":"96c8168c-bf7e-4307-b30c-f4e5dfbebdfe","Type":"ContainerStarted","Data":"c2eb4c2f699a44838281e874e875c30f3abc34505b4619a14bbdb0e801f6d968"} Oct 02 09:57:43 crc kubenswrapper[4771]: I1002 09:57:43.089207 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5556b8d8b5-fj9fv" podStartSLOduration=14.089184344 podStartE2EDuration="14.089184344s" podCreationTimestamp="2025-10-02 09:57:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:57:43.082613756 +0000 UTC m=+1250.730298833" watchObservedRunningTime="2025-10-02 09:57:43.089184344 +0000 UTC m=+1250.736869411" Oct 02 09:57:49 crc kubenswrapper[4771]: E1002 09:57:49.473258 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 02 09:57:49 crc kubenswrapper[4771]: E1002 09:57:49.474770 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qnslr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(ebc075f5-e2ba-4789-a38e-4065fe661ada): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:57:49 crc kubenswrapper[4771]: E1002 09:57:49.476073 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="ebc075f5-e2ba-4789-a38e-4065fe661ada" Oct 02 09:57:49 crc kubenswrapper[4771]: E1002 09:57:49.497503 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 02 09:57:49 crc kubenswrapper[4771]: E1002 09:57:49.497784 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4hz5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(b7dad196-4ebc-43e0-b850-0820f2b072c9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:57:49 crc kubenswrapper[4771]: E1002 09:57:49.499028 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b7dad196-4ebc-43e0-b850-0820f2b072c9" Oct 02 09:57:50 crc kubenswrapper[4771]: E1002 09:57:50.160956 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="ebc075f5-e2ba-4789-a38e-4065fe661ada" Oct 02 09:57:50 crc kubenswrapper[4771]: E1002 09:57:50.162045 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b7dad196-4ebc-43e0-b850-0820f2b072c9" Oct 02 09:57:50 crc kubenswrapper[4771]: I1002 09:57:50.347824 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:50 crc kubenswrapper[4771]: I1002 09:57:50.347879 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:50 crc kubenswrapper[4771]: I1002 09:57:50.355177 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:51 crc kubenswrapper[4771]: I1002 09:57:51.171977 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5556b8d8b5-fj9fv" Oct 02 09:57:51 crc kubenswrapper[4771]: I1002 09:57:51.244749 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-569495bbfd-vb9kq"] Oct 02 09:57:56 crc kubenswrapper[4771]: E1002 09:57:56.611062 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0" Oct 02 09:57:56 crc kubenswrapper[4771]: E1002 09:57:56.611764 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init-config-reloader,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0,Command:[/bin/prometheus-config-reloader],Args:[--watch-interval=0 --listen-address=:8081 --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-metric-storage-rulefiles-0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:reloader-web,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:SHARD,Value:0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/prometheus/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-out,ReadOnly:false,MountPath:/etc/prometheus/config_out,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:prometheus-metric-storage-rulefiles-0,ReadOnly:false,MountPath:/etc/prometheus/rules/prometheus-metric-storage-rulefiles-0,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lknt7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod prometheus-metric-storage-0_openstack(95cd3d5a-1000-4212-b4d9-5d0da549c7f3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:57:56 crc kubenswrapper[4771]: E1002 09:57:56.613800 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-config-reloader\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/prometheus-metric-storage-0" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" Oct 02 09:57:57 crc kubenswrapper[4771]: E1002 09:57:57.236488 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-config-reloader\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0\\\"\"" pod="openstack/prometheus-metric-storage-0" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" Oct 02 09:57:57 crc kubenswrapper[4771]: E1002 09:57:57.393694 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Oct 02 09:57:57 crc kubenswrapper[4771]: E1002 09:57:57.393935 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n68dh558h5c8h5dfh65h9bh76h57ch5h66dh658h55ch579h5cdh8dhb4h588h56bh98h9ch554hd6h5bch76h557hddh679h77hf6h657h644h666q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tvb24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(785673c9-400c-474f-915d-30e8da38b311): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:57:57 crc kubenswrapper[4771]: E1002 09:57:57.395195 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="785673c9-400c-474f-915d-30e8da38b311" Oct 02 09:57:58 crc kubenswrapper[4771]: I1002 09:57:58.037832 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 09:57:58 crc kubenswrapper[4771]: E1002 09:57:58.248957 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="785673c9-400c-474f-915d-30e8da38b311" Oct 02 09:57:58 crc kubenswrapper[4771]: E1002 09:57:58.822771 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 09:57:58 crc kubenswrapper[4771]: E1002 09:57:58.823003 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rspxq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-2v2x2_openstack(989a5396-f3cd-4231-a14d-fe7bac16db35): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:57:58 crc kubenswrapper[4771]: E1002 09:57:58.824231 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" podUID="989a5396-f3cd-4231-a14d-fe7bac16db35" Oct 02 09:57:58 crc kubenswrapper[4771]: E1002 09:57:58.891167 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 09:57:58 crc kubenswrapper[4771]: E1002 09:57:58.891954 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mczfg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-hbd2n_openstack(8a4bd248-50ba-48c2-8425-bd5e86b0e594): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:57:58 crc kubenswrapper[4771]: E1002 09:57:58.893085 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-hbd2n" podUID="8a4bd248-50ba-48c2-8425-bd5e86b0e594" Oct 02 09:57:58 crc kubenswrapper[4771]: W1002 09:57:58.897057 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfd7c23d_a847_44c8_8e36_01ed1520831d.slice/crio-f4a4c635f53beeda84f594439bb422b88826587917d04a382992a2729f590d27 WatchSource:0}: Error finding container f4a4c635f53beeda84f594439bb422b88826587917d04a382992a2729f590d27: Status 404 returned error can't find the container with id f4a4c635f53beeda84f594439bb422b88826587917d04a382992a2729f590d27 Oct 02 09:57:58 crc kubenswrapper[4771]: E1002 09:57:58.910811 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 09:57:58 crc kubenswrapper[4771]: E1002 09:57:58.910975 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9w474,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-7hclw_openstack(4265ff2d-9844-4d4f-922a-efe637c14e8d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:57:58 crc kubenswrapper[4771]: E1002 09:57:58.912098 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" podUID="4265ff2d-9844-4d4f-922a-efe637c14e8d" Oct 02 09:57:59 crc kubenswrapper[4771]: E1002 09:57:59.033198 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 09:57:59 crc kubenswrapper[4771]: E1002 09:57:59.033417 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x8jmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-8h9mn_openstack(0091b0aa-e43f-476c-8383-c3d279dc11d2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:57:59 crc kubenswrapper[4771]: E1002 09:57:59.035961 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" podUID="0091b0aa-e43f-476c-8383-c3d279dc11d2" Oct 02 09:57:59 crc kubenswrapper[4771]: I1002 09:57:59.261742 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cfd7c23d-a847-44c8-8e36-01ed1520831d","Type":"ContainerStarted","Data":"f4a4c635f53beeda84f594439bb422b88826587917d04a382992a2729f590d27"} Oct 02 09:57:59 crc kubenswrapper[4771]: E1002 09:57:59.273546 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" podUID="0091b0aa-e43f-476c-8383-c3d279dc11d2" Oct 02 09:57:59 crc kubenswrapper[4771]: E1002 09:57:59.274319 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" podUID="4265ff2d-9844-4d4f-922a-efe637c14e8d" Oct 02 09:57:59 crc kubenswrapper[4771]: I1002 09:57:59.567748 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5jf58"] Oct 02 09:57:59 crc kubenswrapper[4771]: I1002 09:57:59.617657 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-9qhq9"] Oct 02 09:57:59 crc kubenswrapper[4771]: W1002 09:57:59.931707 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3a8d6f9_c114_49ea_af29_8399e6a598cc.slice/crio-9c883667a90b957d458fe579c159596c42bfab1515a8f2c9626b21d52fcb850d WatchSource:0}: Error finding container 9c883667a90b957d458fe579c159596c42bfab1515a8f2c9626b21d52fcb850d: Status 404 returned error can't find the container with id 9c883667a90b957d458fe579c159596c42bfab1515a8f2c9626b21d52fcb850d Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.061212 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.198190 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-hbd2n" Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.268859 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a4bd248-50ba-48c2-8425-bd5e86b0e594-config\") pod \"8a4bd248-50ba-48c2-8425-bd5e86b0e594\" (UID: \"8a4bd248-50ba-48c2-8425-bd5e86b0e594\") " Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.269217 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mczfg\" (UniqueName: \"kubernetes.io/projected/8a4bd248-50ba-48c2-8425-bd5e86b0e594-kube-api-access-mczfg\") pod \"8a4bd248-50ba-48c2-8425-bd5e86b0e594\" (UID: \"8a4bd248-50ba-48c2-8425-bd5e86b0e594\") " Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.269446 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a4bd248-50ba-48c2-8425-bd5e86b0e594-config" (OuterVolumeSpecName: "config") pod "8a4bd248-50ba-48c2-8425-bd5e86b0e594" (UID: "8a4bd248-50ba-48c2-8425-bd5e86b0e594"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.270409 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a4bd248-50ba-48c2-8425-bd5e86b0e594-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.286785 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-hbd2n" event={"ID":"8a4bd248-50ba-48c2-8425-bd5e86b0e594","Type":"ContainerDied","Data":"5bbb43c5f1202e8eb22328cbe9556acc1e6ee143846b3ab03d3e6acfd87a2896"} Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.287111 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-hbd2n" Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.291969 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5jf58" event={"ID":"3388e290-ed26-4869-b6ce-c93670e7d9cf","Type":"ContainerStarted","Data":"412f7d9b55b03d76c2f246c29ca5568d98329e6e61b7028fc61c9fbac1f1026e"} Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.293572 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9qhq9" event={"ID":"f3a8d6f9-c114-49ea-af29-8399e6a598cc","Type":"ContainerStarted","Data":"9c883667a90b957d458fe579c159596c42bfab1515a8f2c9626b21d52fcb850d"} Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.297067 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a4bd248-50ba-48c2-8425-bd5e86b0e594-kube-api-access-mczfg" (OuterVolumeSpecName: "kube-api-access-mczfg") pod "8a4bd248-50ba-48c2-8425-bd5e86b0e594" (UID: "8a4bd248-50ba-48c2-8425-bd5e86b0e594"). InnerVolumeSpecName "kube-api-access-mczfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.342404 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.371599 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/989a5396-f3cd-4231-a14d-fe7bac16db35-config\") pod \"989a5396-f3cd-4231-a14d-fe7bac16db35\" (UID: \"989a5396-f3cd-4231-a14d-fe7bac16db35\") " Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.371726 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rspxq\" (UniqueName: \"kubernetes.io/projected/989a5396-f3cd-4231-a14d-fe7bac16db35-kube-api-access-rspxq\") pod \"989a5396-f3cd-4231-a14d-fe7bac16db35\" (UID: \"989a5396-f3cd-4231-a14d-fe7bac16db35\") " Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.371818 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/989a5396-f3cd-4231-a14d-fe7bac16db35-dns-svc\") pod \"989a5396-f3cd-4231-a14d-fe7bac16db35\" (UID: \"989a5396-f3cd-4231-a14d-fe7bac16db35\") " Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.372373 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/989a5396-f3cd-4231-a14d-fe7bac16db35-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "989a5396-f3cd-4231-a14d-fe7bac16db35" (UID: "989a5396-f3cd-4231-a14d-fe7bac16db35"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.372471 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/989a5396-f3cd-4231-a14d-fe7bac16db35-config" (OuterVolumeSpecName: "config") pod "989a5396-f3cd-4231-a14d-fe7bac16db35" (UID: "989a5396-f3cd-4231-a14d-fe7bac16db35"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.372574 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mczfg\" (UniqueName: \"kubernetes.io/projected/8a4bd248-50ba-48c2-8425-bd5e86b0e594-kube-api-access-mczfg\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.372594 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/989a5396-f3cd-4231-a14d-fe7bac16db35-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.376991 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/989a5396-f3cd-4231-a14d-fe7bac16db35-kube-api-access-rspxq" (OuterVolumeSpecName: "kube-api-access-rspxq") pod "989a5396-f3cd-4231-a14d-fe7bac16db35" (UID: "989a5396-f3cd-4231-a14d-fe7bac16db35"). InnerVolumeSpecName "kube-api-access-rspxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.475421 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/989a5396-f3cd-4231-a14d-fe7bac16db35-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.475573 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rspxq\" (UniqueName: \"kubernetes.io/projected/989a5396-f3cd-4231-a14d-fe7bac16db35-kube-api-access-rspxq\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.664084 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-hbd2n"] Oct 02 09:58:00 crc kubenswrapper[4771]: I1002 09:58:00.672172 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-hbd2n"] Oct 02 09:58:01 crc kubenswrapper[4771]: I1002 09:58:01.325851 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq" event={"ID":"7636c0e3-830f-4f22-88ca-52a4e1358e08","Type":"ContainerStarted","Data":"dfbb771a62f64d55ad60a8283e88f0314a1c3b37847a3addbdecbef3a50b79b0"} Oct 02 09:58:01 crc kubenswrapper[4771]: I1002 09:58:01.329333 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"795a79a7-e29c-496f-87c0-0bc8a1ef1570","Type":"ContainerStarted","Data":"3eb20de613e9de40853a088b91a6750c36670bad563808e1432091de2e6d1e45"} Oct 02 09:58:01 crc kubenswrapper[4771]: I1002 09:58:01.334910 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"36a32369-c748-4356-be1b-5008d4f9bbe3","Type":"ContainerStarted","Data":"0a2666ca48394ab070ff7c58363eff654909e02b34ff909957226c82c1a5e34a"} Oct 02 09:58:01 crc kubenswrapper[4771]: I1002 09:58:01.338525 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" Oct 02 09:58:01 crc kubenswrapper[4771]: I1002 09:58:01.338555 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2v2x2" event={"ID":"989a5396-f3cd-4231-a14d-fe7bac16db35","Type":"ContainerDied","Data":"9fa7f01913676fb4009f8f3bb9bf2c9917d4d8d15ba842a41de3f6df32495421"} Oct 02 09:58:01 crc kubenswrapper[4771]: I1002 09:58:01.387931 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-6584dc9448-9vbgq" podStartSLOduration=15.213263987 podStartE2EDuration="32.387906436s" podCreationTimestamp="2025-10-02 09:57:29 +0000 UTC" firstStartedPulling="2025-10-02 09:57:41.951877782 +0000 UTC m=+1249.599562849" lastFinishedPulling="2025-10-02 09:57:59.126520221 +0000 UTC m=+1266.774205298" observedRunningTime="2025-10-02 09:58:01.350292495 +0000 UTC m=+1268.997977582" watchObservedRunningTime="2025-10-02 09:58:01.387906436 +0000 UTC m=+1269.035591503" Oct 02 09:58:01 crc kubenswrapper[4771]: I1002 09:58:01.444590 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2v2x2"] Oct 02 09:58:01 crc kubenswrapper[4771]: I1002 09:58:01.454240 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2v2x2"] Oct 02 09:58:01 crc kubenswrapper[4771]: I1002 09:58:01.701845 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a4bd248-50ba-48c2-8425-bd5e86b0e594" path="/var/lib/kubelet/pods/8a4bd248-50ba-48c2-8425-bd5e86b0e594/volumes" Oct 02 09:58:01 crc kubenswrapper[4771]: I1002 09:58:01.702698 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="989a5396-f3cd-4231-a14d-fe7bac16db35" path="/var/lib/kubelet/pods/989a5396-f3cd-4231-a14d-fe7bac16db35/volumes" Oct 02 09:58:02 crc kubenswrapper[4771]: I1002 09:58:02.356983 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"774a9026-9330-4c75-a2ca-33d5bccaf388","Type":"ContainerStarted","Data":"4bb2f537aa84de56c560b29a045547411ebf29dc877aa136c1df84ac887a6f5b"} Oct 02 09:58:02 crc kubenswrapper[4771]: I1002 09:58:02.372101 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d8f54e64-697d-41b5-83fd-295ce5143d04","Type":"ContainerStarted","Data":"d08794f587f9bdf237349764132dd957e40260917b09698e06252abe9c0e6ae9"} Oct 02 09:58:02 crc kubenswrapper[4771]: I1002 09:58:02.457905 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.346444513 podStartE2EDuration="34.457862976s" podCreationTimestamp="2025-10-02 09:57:28 +0000 UTC" firstStartedPulling="2025-10-02 09:57:30.503699323 +0000 UTC m=+1238.151384390" lastFinishedPulling="2025-10-02 09:58:01.615117796 +0000 UTC m=+1269.262802853" observedRunningTime="2025-10-02 09:58:02.406506573 +0000 UTC m=+1270.054191660" watchObservedRunningTime="2025-10-02 09:58:02.457862976 +0000 UTC m=+1270.105548043" Oct 02 09:58:03 crc kubenswrapper[4771]: I1002 09:58:03.402073 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 09:58:04 crc kubenswrapper[4771]: I1002 09:58:04.416059 4771 generic.go:334] "Generic (PLEG): container finished" podID="36a32369-c748-4356-be1b-5008d4f9bbe3" containerID="0a2666ca48394ab070ff7c58363eff654909e02b34ff909957226c82c1a5e34a" exitCode=0 Oct 02 09:58:04 crc kubenswrapper[4771]: I1002 09:58:04.416185 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"36a32369-c748-4356-be1b-5008d4f9bbe3","Type":"ContainerDied","Data":"0a2666ca48394ab070ff7c58363eff654909e02b34ff909957226c82c1a5e34a"} Oct 02 09:58:06 crc kubenswrapper[4771]: I1002 09:58:06.437183 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5jf58" event={"ID":"3388e290-ed26-4869-b6ce-c93670e7d9cf","Type":"ContainerStarted","Data":"fb47923a18fdefe0144f61113597d9b92a1170935abdaecf65fec848b2c9bf84"} Oct 02 09:58:06 crc kubenswrapper[4771]: I1002 09:58:06.439024 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-5jf58" Oct 02 09:58:06 crc kubenswrapper[4771]: I1002 09:58:06.440625 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9qhq9" event={"ID":"f3a8d6f9-c114-49ea-af29-8399e6a598cc","Type":"ContainerStarted","Data":"fea41e492d14894f300c75ccb8fa503e71b9bae4eea1d6461680ebf94668992f"} Oct 02 09:58:06 crc kubenswrapper[4771]: I1002 09:58:06.443404 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cfd7c23d-a847-44c8-8e36-01ed1520831d","Type":"ContainerStarted","Data":"ccb472c955434c86da4d03f3261f18fc7a5d0ef7346797956b6bab78bf34f37d"} Oct 02 09:58:06 crc kubenswrapper[4771]: I1002 09:58:06.445259 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"795a79a7-e29c-496f-87c0-0bc8a1ef1570","Type":"ContainerStarted","Data":"7a1f5aed2a8395803523c6404d8487d67a08e5438a68bc1485ba017604aabc8a"} Oct 02 09:58:06 crc kubenswrapper[4771]: I1002 09:58:06.446965 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"36a32369-c748-4356-be1b-5008d4f9bbe3","Type":"ContainerStarted","Data":"3103211da5e839cf61bb40c800b0f50ab71df069469e8b7c897f89020439a2cc"} Oct 02 09:58:06 crc kubenswrapper[4771]: I1002 09:58:06.448861 4771 generic.go:334] "Generic (PLEG): container finished" podID="774a9026-9330-4c75-a2ca-33d5bccaf388" containerID="4bb2f537aa84de56c560b29a045547411ebf29dc877aa136c1df84ac887a6f5b" exitCode=0 Oct 02 09:58:06 crc kubenswrapper[4771]: I1002 09:58:06.448895 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"774a9026-9330-4c75-a2ca-33d5bccaf388","Type":"ContainerDied","Data":"4bb2f537aa84de56c560b29a045547411ebf29dc877aa136c1df84ac887a6f5b"} Oct 02 09:58:06 crc kubenswrapper[4771]: I1002 09:58:06.460272 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-5jf58" podStartSLOduration=28.951634891 podStartE2EDuration="34.46025169s" podCreationTimestamp="2025-10-02 09:57:32 +0000 UTC" firstStartedPulling="2025-10-02 09:58:00.088926371 +0000 UTC m=+1267.736611438" lastFinishedPulling="2025-10-02 09:58:05.59754317 +0000 UTC m=+1273.245228237" observedRunningTime="2025-10-02 09:58:06.459815289 +0000 UTC m=+1274.107500356" watchObservedRunningTime="2025-10-02 09:58:06.46025169 +0000 UTC m=+1274.107936757" Oct 02 09:58:06 crc kubenswrapper[4771]: I1002 09:58:06.490531 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=11.655406858 podStartE2EDuration="41.490512474s" podCreationTimestamp="2025-10-02 09:57:25 +0000 UTC" firstStartedPulling="2025-10-02 09:57:28.028598333 +0000 UTC m=+1235.676283410" lastFinishedPulling="2025-10-02 09:57:57.863703959 +0000 UTC m=+1265.511389026" observedRunningTime="2025-10-02 09:58:06.481249687 +0000 UTC m=+1274.128934774" watchObservedRunningTime="2025-10-02 09:58:06.490512474 +0000 UTC m=+1274.138197541" Oct 02 09:58:06 crc kubenswrapper[4771]: I1002 09:58:06.858124 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 02 09:58:06 crc kubenswrapper[4771]: I1002 09:58:06.858488 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 02 09:58:07 crc kubenswrapper[4771]: I1002 09:58:07.462293 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"774a9026-9330-4c75-a2ca-33d5bccaf388","Type":"ContainerStarted","Data":"15827cde2da486edce617a63d64ffbe554604f3fb53213556b488f57c921e970"} Oct 02 09:58:07 crc kubenswrapper[4771]: I1002 09:58:07.466280 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7dad196-4ebc-43e0-b850-0820f2b072c9","Type":"ContainerStarted","Data":"63aadd758b05567dc1365ba3e674b8ba28ce7f2c61ec9a5fcd81f0d9682d8a7c"} Oct 02 09:58:07 crc kubenswrapper[4771]: I1002 09:58:07.469240 4771 generic.go:334] "Generic (PLEG): container finished" podID="f3a8d6f9-c114-49ea-af29-8399e6a598cc" containerID="fea41e492d14894f300c75ccb8fa503e71b9bae4eea1d6461680ebf94668992f" exitCode=0 Oct 02 09:58:07 crc kubenswrapper[4771]: I1002 09:58:07.469316 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9qhq9" event={"ID":"f3a8d6f9-c114-49ea-af29-8399e6a598cc","Type":"ContainerDied","Data":"fea41e492d14894f300c75ccb8fa503e71b9bae4eea1d6461680ebf94668992f"} Oct 02 09:58:07 crc kubenswrapper[4771]: I1002 09:58:07.472936 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ebc075f5-e2ba-4789-a38e-4065fe661ada","Type":"ContainerStarted","Data":"74e0addf34252d1465b8adca2fc82ba42dcd867d6b03aac7aec4c69a2f14976d"} Oct 02 09:58:07 crc kubenswrapper[4771]: I1002 09:58:07.574567 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=12.068373986 podStartE2EDuration="44.574537424s" podCreationTimestamp="2025-10-02 09:57:23 +0000 UTC" firstStartedPulling="2025-10-02 09:57:26.619027259 +0000 UTC m=+1234.266712336" lastFinishedPulling="2025-10-02 09:57:59.125190707 +0000 UTC m=+1266.772875774" observedRunningTime="2025-10-02 09:58:07.48796512 +0000 UTC m=+1275.135650187" watchObservedRunningTime="2025-10-02 09:58:07.574537424 +0000 UTC m=+1275.222222491" Oct 02 09:58:08 crc kubenswrapper[4771]: I1002 09:58:08.496036 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9qhq9" event={"ID":"f3a8d6f9-c114-49ea-af29-8399e6a598cc","Type":"ContainerStarted","Data":"a94b98c6cac0a6a266778279b01e4f4def45bebe021ae66d83d75110d12aecba"} Oct 02 09:58:08 crc kubenswrapper[4771]: I1002 09:58:08.496686 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9qhq9" event={"ID":"f3a8d6f9-c114-49ea-af29-8399e6a598cc","Type":"ContainerStarted","Data":"2924364a6596284eb11359c193757d203aa4b2380ad905df50200598e338a0b7"} Oct 02 09:58:08 crc kubenswrapper[4771]: I1002 09:58:08.496746 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:58:08 crc kubenswrapper[4771]: I1002 09:58:08.529713 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-9qhq9" podStartSLOduration=30.876881031 podStartE2EDuration="36.529685607s" podCreationTimestamp="2025-10-02 09:57:32 +0000 UTC" firstStartedPulling="2025-10-02 09:57:59.943236775 +0000 UTC m=+1267.590921842" lastFinishedPulling="2025-10-02 09:58:05.596041351 +0000 UTC m=+1273.243726418" observedRunningTime="2025-10-02 09:58:08.519994499 +0000 UTC m=+1276.167679576" watchObservedRunningTime="2025-10-02 09:58:08.529685607 +0000 UTC m=+1276.177370674" Oct 02 09:58:08 crc kubenswrapper[4771]: I1002 09:58:08.683907 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 09:58:09 crc kubenswrapper[4771]: I1002 09:58:09.506821 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:58:15 crc kubenswrapper[4771]: I1002 09:58:15.310663 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 02 09:58:15 crc kubenswrapper[4771]: I1002 09:58:15.311315 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 02 09:58:15 crc kubenswrapper[4771]: I1002 09:58:15.638531 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 02 09:58:15 crc kubenswrapper[4771]: I1002 09:58:15.740945 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="36a32369-c748-4356-be1b-5008d4f9bbe3" containerName="galera" probeResult="failure" output=< Oct 02 09:58:15 crc kubenswrapper[4771]: wsrep_local_state_comment (Joined) differs from Synced Oct 02 09:58:15 crc kubenswrapper[4771]: > Oct 02 09:58:16 crc kubenswrapper[4771]: I1002 09:58:16.329393 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-569495bbfd-vb9kq" podUID="e5be9e93-f0cb-41e5-b232-6f02e4337d59" containerName="console" containerID="cri-o://0ad6c8f8a8cc4b33f9a2647b4b650bba50991ccc352407fa684e96de80d8cad6" gracePeriod=15 Oct 02 09:58:16 crc kubenswrapper[4771]: I1002 09:58:16.586947 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-569495bbfd-vb9kq_e5be9e93-f0cb-41e5-b232-6f02e4337d59/console/0.log" Oct 02 09:58:16 crc kubenswrapper[4771]: I1002 09:58:16.587283 4771 generic.go:334] "Generic (PLEG): container finished" podID="e5be9e93-f0cb-41e5-b232-6f02e4337d59" containerID="0ad6c8f8a8cc4b33f9a2647b4b650bba50991ccc352407fa684e96de80d8cad6" exitCode=2 Oct 02 09:58:16 crc kubenswrapper[4771]: I1002 09:58:16.587323 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-569495bbfd-vb9kq" event={"ID":"e5be9e93-f0cb-41e5-b232-6f02e4337d59","Type":"ContainerDied","Data":"0ad6c8f8a8cc4b33f9a2647b4b650bba50991ccc352407fa684e96de80d8cad6"} Oct 02 09:58:16 crc kubenswrapper[4771]: I1002 09:58:16.911332 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 02 09:58:21 crc kubenswrapper[4771]: E1002 09:58:21.705217 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Oct 02 09:58:21 crc kubenswrapper[4771]: E1002 09:58:21.705823 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ssz2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(795a79a7-e29c-496f-87c0-0bc8a1ef1570): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:58:21 crc kubenswrapper[4771]: E1002 09:58:21.707029 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="795a79a7-e29c-496f-87c0-0bc8a1ef1570" Oct 02 09:58:21 crc kubenswrapper[4771]: I1002 09:58:21.849403 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 02 09:58:21 crc kubenswrapper[4771]: I1002 09:58:21.939534 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="774a9026-9330-4c75-a2ca-33d5bccaf388" containerName="galera" probeResult="failure" output=< Oct 02 09:58:21 crc kubenswrapper[4771]: wsrep_local_state_comment (Joined) differs from Synced Oct 02 09:58:21 crc kubenswrapper[4771]: > Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.498163 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-569495bbfd-vb9kq_e5be9e93-f0cb-41e5-b232-6f02e4337d59/console/0.log" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.498648 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.580143 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-oauth-config\") pod \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.580197 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgplb\" (UniqueName: \"kubernetes.io/projected/e5be9e93-f0cb-41e5-b232-6f02e4337d59-kube-api-access-bgplb\") pod \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.580243 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-service-ca\") pod \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.580322 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-oauth-serving-cert\") pod \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.580338 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-trusted-ca-bundle\") pod \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.580449 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-config\") pod \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.580464 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-serving-cert\") pod \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\" (UID: \"e5be9e93-f0cb-41e5-b232-6f02e4337d59\") " Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.581331 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-service-ca" (OuterVolumeSpecName: "service-ca") pod "e5be9e93-f0cb-41e5-b232-6f02e4337d59" (UID: "e5be9e93-f0cb-41e5-b232-6f02e4337d59"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.581423 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "e5be9e93-f0cb-41e5-b232-6f02e4337d59" (UID: "e5be9e93-f0cb-41e5-b232-6f02e4337d59"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.581464 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-config" (OuterVolumeSpecName: "console-config") pod "e5be9e93-f0cb-41e5-b232-6f02e4337d59" (UID: "e5be9e93-f0cb-41e5-b232-6f02e4337d59"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.581616 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "e5be9e93-f0cb-41e5-b232-6f02e4337d59" (UID: "e5be9e93-f0cb-41e5-b232-6f02e4337d59"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.585963 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "e5be9e93-f0cb-41e5-b232-6f02e4337d59" (UID: "e5be9e93-f0cb-41e5-b232-6f02e4337d59"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.586092 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "e5be9e93-f0cb-41e5-b232-6f02e4337d59" (UID: "e5be9e93-f0cb-41e5-b232-6f02e4337d59"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.586639 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5be9e93-f0cb-41e5-b232-6f02e4337d59-kube-api-access-bgplb" (OuterVolumeSpecName: "kube-api-access-bgplb") pod "e5be9e93-f0cb-41e5-b232-6f02e4337d59" (UID: "e5be9e93-f0cb-41e5-b232-6f02e4337d59"). InnerVolumeSpecName "kube-api-access-bgplb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.669283 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-569495bbfd-vb9kq_e5be9e93-f0cb-41e5-b232-6f02e4337d59/console/0.log" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.669358 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-569495bbfd-vb9kq" event={"ID":"e5be9e93-f0cb-41e5-b232-6f02e4337d59","Type":"ContainerDied","Data":"a6f2fa9bcb3240d7aadc4dffab955b869f2b3477a670200c221ae903a32c661c"} Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.669410 4771 scope.go:117] "RemoveContainer" containerID="0ad6c8f8a8cc4b33f9a2647b4b650bba50991ccc352407fa684e96de80d8cad6" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.669503 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-569495bbfd-vb9kq" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.678807 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.684205 4771 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.684258 4771 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.684271 4771 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.684280 4771 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.684307 4771 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e5be9e93-f0cb-41e5-b232-6f02e4337d59-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.684317 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgplb\" (UniqueName: \"kubernetes.io/projected/e5be9e93-f0cb-41e5-b232-6f02e4337d59-kube-api-access-bgplb\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.684329 4771 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e5be9e93-f0cb-41e5-b232-6f02e4337d59-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.724226 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-569495bbfd-vb9kq"] Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.730883 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-569495bbfd-vb9kq"] Oct 02 09:58:23 crc kubenswrapper[4771]: I1002 09:58:23.767558 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 02 09:58:24 crc kubenswrapper[4771]: I1002 09:58:24.683718 4771 generic.go:334] "Generic (PLEG): container finished" podID="0091b0aa-e43f-476c-8383-c3d279dc11d2" containerID="841ad4e2563abbb9c7417b30557ee4a224a2f5f20713a6e3af69c1cad31829f6" exitCode=0 Oct 02 09:58:24 crc kubenswrapper[4771]: I1002 09:58:24.683843 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" event={"ID":"0091b0aa-e43f-476c-8383-c3d279dc11d2","Type":"ContainerDied","Data":"841ad4e2563abbb9c7417b30557ee4a224a2f5f20713a6e3af69c1cad31829f6"} Oct 02 09:58:24 crc kubenswrapper[4771]: I1002 09:58:24.688506 4771 generic.go:334] "Generic (PLEG): container finished" podID="4265ff2d-9844-4d4f-922a-efe637c14e8d" containerID="4f8147b79e28ceeee6d5ed5931c7135f3876fa2bf3de5b81d07dd2f380427504" exitCode=0 Oct 02 09:58:24 crc kubenswrapper[4771]: I1002 09:58:24.688585 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" event={"ID":"4265ff2d-9844-4d4f-922a-efe637c14e8d","Type":"ContainerDied","Data":"4f8147b79e28ceeee6d5ed5931c7135f3876fa2bf3de5b81d07dd2f380427504"} Oct 02 09:58:24 crc kubenswrapper[4771]: I1002 09:58:24.691784 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"795a79a7-e29c-496f-87c0-0bc8a1ef1570","Type":"ContainerStarted","Data":"bae9f6895ab71ec4899c384663fec6877c3d90f52d0fed9cc912f37b432f8b2b"} Oct 02 09:58:24 crc kubenswrapper[4771]: I1002 09:58:24.692642 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 02 09:58:24 crc kubenswrapper[4771]: I1002 09:58:24.697017 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"785673c9-400c-474f-915d-30e8da38b311","Type":"ContainerStarted","Data":"88d76340526845f3b6d0087b6120ba8bc7afab23ee8bebeff91549d240f23ee7"} Oct 02 09:58:24 crc kubenswrapper[4771]: I1002 09:58:24.697261 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 02 09:58:24 crc kubenswrapper[4771]: I1002 09:58:24.704505 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cfd7c23d-a847-44c8-8e36-01ed1520831d","Type":"ContainerStarted","Data":"fcd220d30008136d904c65913489bf57ce1bd753207d59126f1061a3f877e67d"} Oct 02 09:58:24 crc kubenswrapper[4771]: I1002 09:58:24.729627 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.881087652 podStartE2EDuration="58.729600392s" podCreationTimestamp="2025-10-02 09:57:26 +0000 UTC" firstStartedPulling="2025-10-02 09:57:27.921295889 +0000 UTC m=+1235.568980956" lastFinishedPulling="2025-10-02 09:58:23.769808629 +0000 UTC m=+1291.417493696" observedRunningTime="2025-10-02 09:58:24.717760719 +0000 UTC m=+1292.365445806" watchObservedRunningTime="2025-10-02 09:58:24.729600392 +0000 UTC m=+1292.377285469" Oct 02 09:58:24 crc kubenswrapper[4771]: I1002 09:58:24.749984 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=45.460044446 podStartE2EDuration="50.749959083s" podCreationTimestamp="2025-10-02 09:57:34 +0000 UTC" firstStartedPulling="2025-10-02 09:58:00.308885305 +0000 UTC m=+1267.956570372" lastFinishedPulling="2025-10-02 09:58:05.598799942 +0000 UTC m=+1273.246485009" observedRunningTime="2025-10-02 09:58:24.746543705 +0000 UTC m=+1292.394228962" watchObservedRunningTime="2025-10-02 09:58:24.749959083 +0000 UTC m=+1292.397644150" Oct 02 09:58:24 crc kubenswrapper[4771]: I1002 09:58:24.754407 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 02 09:58:24 crc kubenswrapper[4771]: I1002 09:58:24.831278 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=29.749132592 podStartE2EDuration="52.831250871s" podCreationTimestamp="2025-10-02 09:57:32 +0000 UTC" firstStartedPulling="2025-10-02 09:57:58.910816285 +0000 UTC m=+1266.558501352" lastFinishedPulling="2025-10-02 09:58:21.992934564 +0000 UTC m=+1289.640619631" observedRunningTime="2025-10-02 09:58:24.822673672 +0000 UTC m=+1292.470358739" watchObservedRunningTime="2025-10-02 09:58:24.831250871 +0000 UTC m=+1292.478935938" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.010919 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-7hclw"] Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.045733 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-vcdhf"] Oct 02 09:58:25 crc kubenswrapper[4771]: E1002 09:58:25.046264 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5be9e93-f0cb-41e5-b232-6f02e4337d59" containerName="console" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.046286 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5be9e93-f0cb-41e5-b232-6f02e4337d59" containerName="console" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.046570 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5be9e93-f0cb-41e5-b232-6f02e4337d59" containerName="console" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.068979 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.082510 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.087635 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-vcdhf"] Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.108691 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-x9dc5"] Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.110420 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.114435 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.116391 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-vcdhf\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.116492 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxkr2\" (UniqueName: \"kubernetes.io/projected/ff25c029-7927-45c7-bc00-bc1c118cb8fd-kube-api-access-pxkr2\") pod \"dnsmasq-dns-7f896c8c65-vcdhf\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.116539 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-config\") pod \"dnsmasq-dns-7f896c8c65-vcdhf\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.116578 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-vcdhf\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.127722 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-x9dc5"] Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.224399 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dmc8\" (UniqueName: \"kubernetes.io/projected/7e2e2fc3-4cf2-4455-a687-c07b61aed147-kube-api-access-7dmc8\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.224456 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxkr2\" (UniqueName: \"kubernetes.io/projected/ff25c029-7927-45c7-bc00-bc1c118cb8fd-kube-api-access-pxkr2\") pod \"dnsmasq-dns-7f896c8c65-vcdhf\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.224496 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e2e2fc3-4cf2-4455-a687-c07b61aed147-config\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.224521 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-config\") pod \"dnsmasq-dns-7f896c8c65-vcdhf\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.224552 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-vcdhf\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.224609 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e2e2fc3-4cf2-4455-a687-c07b61aed147-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.224643 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2e2fc3-4cf2-4455-a687-c07b61aed147-combined-ca-bundle\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.224753 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7e2e2fc3-4cf2-4455-a687-c07b61aed147-ovn-rundir\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.224783 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-vcdhf\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.224817 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7e2e2fc3-4cf2-4455-a687-c07b61aed147-ovs-rundir\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.225753 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-vcdhf\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.225805 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-config\") pod \"dnsmasq-dns-7f896c8c65-vcdhf\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.226324 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-vcdhf\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.246532 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxkr2\" (UniqueName: \"kubernetes.io/projected/ff25c029-7927-45c7-bc00-bc1c118cb8fd-kube-api-access-pxkr2\") pod \"dnsmasq-dns-7f896c8c65-vcdhf\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.266147 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8h9mn"] Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.295656 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-cwm9n"] Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.297639 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.303982 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.322033 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-cwm9n"] Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.325873 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e2e2fc3-4cf2-4455-a687-c07b61aed147-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.325937 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2e2fc3-4cf2-4455-a687-c07b61aed147-combined-ca-bundle\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.325963 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.325996 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.326023 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7e2e2fc3-4cf2-4455-a687-c07b61aed147-ovn-rundir\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.326072 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7e2e2fc3-4cf2-4455-a687-c07b61aed147-ovs-rundir\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.326123 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dmc8\" (UniqueName: \"kubernetes.io/projected/7e2e2fc3-4cf2-4455-a687-c07b61aed147-kube-api-access-7dmc8\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.326172 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e2e2fc3-4cf2-4455-a687-c07b61aed147-config\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.326189 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.326214 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-config\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.326264 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp9ww\" (UniqueName: \"kubernetes.io/projected/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-kube-api-access-hp9ww\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.326620 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7e2e2fc3-4cf2-4455-a687-c07b61aed147-ovs-rundir\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.326684 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7e2e2fc3-4cf2-4455-a687-c07b61aed147-ovn-rundir\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.327319 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e2e2fc3-4cf2-4455-a687-c07b61aed147-config\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.330847 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e2e2fc3-4cf2-4455-a687-c07b61aed147-combined-ca-bundle\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.345696 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e2e2fc3-4cf2-4455-a687-c07b61aed147-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.353095 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dmc8\" (UniqueName: \"kubernetes.io/projected/7e2e2fc3-4cf2-4455-a687-c07b61aed147-kube-api-access-7dmc8\") pod \"ovn-controller-metrics-x9dc5\" (UID: \"7e2e2fc3-4cf2-4455-a687-c07b61aed147\") " pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.394645 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.428513 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.428568 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-config\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.428629 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp9ww\" (UniqueName: \"kubernetes.io/projected/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-kube-api-access-hp9ww\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.428684 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.428736 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.429713 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-config\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.429734 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.430292 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.430474 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.430494 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.458716 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-x9dc5" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.465216 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp9ww\" (UniqueName: \"kubernetes.io/projected/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-kube-api-access-hp9ww\") pod \"dnsmasq-dns-86db49b7ff-cwm9n\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.630598 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.713148 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5be9e93-f0cb-41e5-b232-6f02e4337d59" path="/var/lib/kubelet/pods/e5be9e93-f0cb-41e5-b232-6f02e4337d59/volumes" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.748932 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" event={"ID":"0091b0aa-e43f-476c-8383-c3d279dc11d2","Type":"ContainerStarted","Data":"867e182039885b01a702368079a7be3979b5981c01662ee61cd415ca8df3eec1"} Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.749111 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" podUID="0091b0aa-e43f-476c-8383-c3d279dc11d2" containerName="dnsmasq-dns" containerID="cri-o://867e182039885b01a702368079a7be3979b5981c01662ee61cd415ca8df3eec1" gracePeriod=10 Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.749238 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.780638 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" podStartSLOduration=3.028746875 podStartE2EDuration="1m3.780613238s" podCreationTimestamp="2025-10-02 09:57:22 +0000 UTC" firstStartedPulling="2025-10-02 09:57:23.40448114 +0000 UTC m=+1231.052166207" lastFinishedPulling="2025-10-02 09:58:24.156347513 +0000 UTC m=+1291.804032570" observedRunningTime="2025-10-02 09:58:25.770978741 +0000 UTC m=+1293.418663808" watchObservedRunningTime="2025-10-02 09:58:25.780613238 +0000 UTC m=+1293.428298325" Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.798229 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" event={"ID":"4265ff2d-9844-4d4f-922a-efe637c14e8d","Type":"ContainerStarted","Data":"9eae86c3847c6698cad06d23de888102374ae3d5ceeb3625d76966b6461733f0"} Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.798466 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" podUID="4265ff2d-9844-4d4f-922a-efe637c14e8d" containerName="dnsmasq-dns" containerID="cri-o://9eae86c3847c6698cad06d23de888102374ae3d5ceeb3625d76966b6461733f0" gracePeriod=10 Oct 02 09:58:25 crc kubenswrapper[4771]: I1002 09:58:25.825602 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" podStartSLOduration=4.213659553 podStartE2EDuration="1m4.825563707s" podCreationTimestamp="2025-10-02 09:57:21 +0000 UTC" firstStartedPulling="2025-10-02 09:57:23.158953942 +0000 UTC m=+1230.806639009" lastFinishedPulling="2025-10-02 09:58:23.770858096 +0000 UTC m=+1291.418543163" observedRunningTime="2025-10-02 09:58:25.822970591 +0000 UTC m=+1293.470655658" watchObservedRunningTime="2025-10-02 09:58:25.825563707 +0000 UTC m=+1293.473248774" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.056516 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-x9dc5"] Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.249098 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-vcdhf"] Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.344843 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-cwm9n"] Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.392695 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-s4xgr"] Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.394343 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s4xgr" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.413927 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-s4xgr"] Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.485940 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr7jn\" (UniqueName: \"kubernetes.io/projected/e87072ca-e829-43db-81d0-80df8e4339a7-kube-api-access-pr7jn\") pod \"keystone-db-create-s4xgr\" (UID: \"e87072ca-e829-43db-81d0-80df8e4339a7\") " pod="openstack/keystone-db-create-s4xgr" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.588550 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-c52xm"] Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.589749 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr7jn\" (UniqueName: \"kubernetes.io/projected/e87072ca-e829-43db-81d0-80df8e4339a7-kube-api-access-pr7jn\") pod \"keystone-db-create-s4xgr\" (UID: \"e87072ca-e829-43db-81d0-80df8e4339a7\") " pod="openstack/keystone-db-create-s4xgr" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.590307 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c52xm" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.602866 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-c52xm"] Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.614499 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr7jn\" (UniqueName: \"kubernetes.io/projected/e87072ca-e829-43db-81d0-80df8e4339a7-kube-api-access-pr7jn\") pod \"keystone-db-create-s4xgr\" (UID: \"e87072ca-e829-43db-81d0-80df8e4339a7\") " pod="openstack/keystone-db-create-s4xgr" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.692038 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xc4j\" (UniqueName: \"kubernetes.io/projected/406c250c-ea6b-4820-b38b-6c4d5c0a16de-kube-api-access-7xc4j\") pod \"placement-db-create-c52xm\" (UID: \"406c250c-ea6b-4820-b38b-6c4d5c0a16de\") " pod="openstack/placement-db-create-c52xm" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.794452 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xc4j\" (UniqueName: \"kubernetes.io/projected/406c250c-ea6b-4820-b38b-6c4d5c0a16de-kube-api-access-7xc4j\") pod \"placement-db-create-c52xm\" (UID: \"406c250c-ea6b-4820-b38b-6c4d5c0a16de\") " pod="openstack/placement-db-create-c52xm" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.808167 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" event={"ID":"ff25c029-7927-45c7-bc00-bc1c118cb8fd","Type":"ContainerStarted","Data":"ecafcbf06a969847abe1fc9ee58557183461028a15dc95123269e76d6ef3da3c"} Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.810671 4771 generic.go:334] "Generic (PLEG): container finished" podID="0091b0aa-e43f-476c-8383-c3d279dc11d2" containerID="867e182039885b01a702368079a7be3979b5981c01662ee61cd415ca8df3eec1" exitCode=0 Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.810790 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" event={"ID":"0091b0aa-e43f-476c-8383-c3d279dc11d2","Type":"ContainerDied","Data":"867e182039885b01a702368079a7be3979b5981c01662ee61cd415ca8df3eec1"} Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.810834 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" event={"ID":"0091b0aa-e43f-476c-8383-c3d279dc11d2","Type":"ContainerDied","Data":"d15415aa27347b5656dc264942fa8e8acfab550e93c74786133a67786249c044"} Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.810851 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d15415aa27347b5656dc264942fa8e8acfab550e93c74786133a67786249c044" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.814682 4771 generic.go:334] "Generic (PLEG): container finished" podID="4265ff2d-9844-4d4f-922a-efe637c14e8d" containerID="9eae86c3847c6698cad06d23de888102374ae3d5ceeb3625d76966b6461733f0" exitCode=0 Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.814809 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" event={"ID":"4265ff2d-9844-4d4f-922a-efe637c14e8d","Type":"ContainerDied","Data":"9eae86c3847c6698cad06d23de888102374ae3d5ceeb3625d76966b6461733f0"} Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.814838 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" event={"ID":"4265ff2d-9844-4d4f-922a-efe637c14e8d","Type":"ContainerDied","Data":"e7b6f882f88ce8ab6df3fd121a5f3ef2f1580273bdef98d635bba07406237587"} Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.814854 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7b6f882f88ce8ab6df3fd121a5f3ef2f1580273bdef98d635bba07406237587" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.816575 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-x9dc5" event={"ID":"7e2e2fc3-4cf2-4455-a687-c07b61aed147","Type":"ContainerStarted","Data":"0b9233ad658072a7755eab04f8fb5e7c2ae2e46c36713815312c8a9233a81457"} Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.818656 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" event={"ID":"7eb14f25-cd51-4f54-ac31-3eb9cf51e152","Type":"ContainerStarted","Data":"0ed8aafc01642f615bb3a2761489092d237108ea4159b07d34e646d67226ad5f"} Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.830661 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xc4j\" (UniqueName: \"kubernetes.io/projected/406c250c-ea6b-4820-b38b-6c4d5c0a16de-kube-api-access-7xc4j\") pod \"placement-db-create-c52xm\" (UID: \"406c250c-ea6b-4820-b38b-6c4d5c0a16de\") " pod="openstack/placement-db-create-c52xm" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.869390 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c52xm" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.880681 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s4xgr" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.896486 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" Oct 02 09:58:26 crc kubenswrapper[4771]: I1002 09:58:26.909935 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.004988 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8jmq\" (UniqueName: \"kubernetes.io/projected/0091b0aa-e43f-476c-8383-c3d279dc11d2-kube-api-access-x8jmq\") pod \"0091b0aa-e43f-476c-8383-c3d279dc11d2\" (UID: \"0091b0aa-e43f-476c-8383-c3d279dc11d2\") " Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.005080 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0091b0aa-e43f-476c-8383-c3d279dc11d2-config\") pod \"0091b0aa-e43f-476c-8383-c3d279dc11d2\" (UID: \"0091b0aa-e43f-476c-8383-c3d279dc11d2\") " Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.005248 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4265ff2d-9844-4d4f-922a-efe637c14e8d-config\") pod \"4265ff2d-9844-4d4f-922a-efe637c14e8d\" (UID: \"4265ff2d-9844-4d4f-922a-efe637c14e8d\") " Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.005290 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0091b0aa-e43f-476c-8383-c3d279dc11d2-dns-svc\") pod \"0091b0aa-e43f-476c-8383-c3d279dc11d2\" (UID: \"0091b0aa-e43f-476c-8383-c3d279dc11d2\") " Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.005356 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w474\" (UniqueName: \"kubernetes.io/projected/4265ff2d-9844-4d4f-922a-efe637c14e8d-kube-api-access-9w474\") pod \"4265ff2d-9844-4d4f-922a-efe637c14e8d\" (UID: \"4265ff2d-9844-4d4f-922a-efe637c14e8d\") " Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.005405 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4265ff2d-9844-4d4f-922a-efe637c14e8d-dns-svc\") pod \"4265ff2d-9844-4d4f-922a-efe637c14e8d\" (UID: \"4265ff2d-9844-4d4f-922a-efe637c14e8d\") " Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.211647 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0091b0aa-e43f-476c-8383-c3d279dc11d2-kube-api-access-x8jmq" (OuterVolumeSpecName: "kube-api-access-x8jmq") pod "0091b0aa-e43f-476c-8383-c3d279dc11d2" (UID: "0091b0aa-e43f-476c-8383-c3d279dc11d2"). InnerVolumeSpecName "kube-api-access-x8jmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.214629 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4265ff2d-9844-4d4f-922a-efe637c14e8d-kube-api-access-9w474" (OuterVolumeSpecName: "kube-api-access-9w474") pod "4265ff2d-9844-4d4f-922a-efe637c14e8d" (UID: "4265ff2d-9844-4d4f-922a-efe637c14e8d"). InnerVolumeSpecName "kube-api-access-9w474". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.315665 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8jmq\" (UniqueName: \"kubernetes.io/projected/0091b0aa-e43f-476c-8383-c3d279dc11d2-kube-api-access-x8jmq\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.315706 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w474\" (UniqueName: \"kubernetes.io/projected/4265ff2d-9844-4d4f-922a-efe637c14e8d-kube-api-access-9w474\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.504557 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4265ff2d-9844-4d4f-922a-efe637c14e8d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4265ff2d-9844-4d4f-922a-efe637c14e8d" (UID: "4265ff2d-9844-4d4f-922a-efe637c14e8d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.522782 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4265ff2d-9844-4d4f-922a-efe637c14e8d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.568685 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-c52xm"] Oct 02 09:58:27 crc kubenswrapper[4771]: W1002 09:58:27.574180 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod406c250c_ea6b_4820_b38b_6c4d5c0a16de.slice/crio-2da22b4c8e4a4df292b96bf0ae6ba91d3dee2c12e4bcc3712d6708b54bd2fafa WatchSource:0}: Error finding container 2da22b4c8e4a4df292b96bf0ae6ba91d3dee2c12e4bcc3712d6708b54bd2fafa: Status 404 returned error can't find the container with id 2da22b4c8e4a4df292b96bf0ae6ba91d3dee2c12e4bcc3712d6708b54bd2fafa Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.606266 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4265ff2d-9844-4d4f-922a-efe637c14e8d-config" (OuterVolumeSpecName: "config") pod "4265ff2d-9844-4d4f-922a-efe637c14e8d" (UID: "4265ff2d-9844-4d4f-922a-efe637c14e8d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.627816 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4265ff2d-9844-4d4f-922a-efe637c14e8d-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.627760 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0091b0aa-e43f-476c-8383-c3d279dc11d2-config" (OuterVolumeSpecName: "config") pod "0091b0aa-e43f-476c-8383-c3d279dc11d2" (UID: "0091b0aa-e43f-476c-8383-c3d279dc11d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.633693 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.696866 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0091b0aa-e43f-476c-8383-c3d279dc11d2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0091b0aa-e43f-476c-8383-c3d279dc11d2" (UID: "0091b0aa-e43f-476c-8383-c3d279dc11d2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.746269 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0091b0aa-e43f-476c-8383-c3d279dc11d2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.746309 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0091b0aa-e43f-476c-8383-c3d279dc11d2-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.760743 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.835053 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"95cd3d5a-1000-4212-b4d9-5d0da549c7f3","Type":"ContainerStarted","Data":"6fd70e48c855a5f9bf81cbf37919a8679ecbde45a91dccd0e2a6fc1a67916fa6"} Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.838527 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-s4xgr"] Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.844704 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-x9dc5" event={"ID":"7e2e2fc3-4cf2-4455-a687-c07b61aed147","Type":"ContainerStarted","Data":"ce8195737b910fcb68a6d91a7300220777bd220ea98e5539ca4a11874bbfc51d"} Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.849597 4771 generic.go:334] "Generic (PLEG): container finished" podID="7eb14f25-cd51-4f54-ac31-3eb9cf51e152" containerID="93ff4861ff36ead6cbbbee3368d13a88352a9bd65803366637c7b6d1a8f68246" exitCode=0 Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.850155 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" event={"ID":"7eb14f25-cd51-4f54-ac31-3eb9cf51e152","Type":"ContainerDied","Data":"93ff4861ff36ead6cbbbee3368d13a88352a9bd65803366637c7b6d1a8f68246"} Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.857446 4771 generic.go:334] "Generic (PLEG): container finished" podID="ff25c029-7927-45c7-bc00-bc1c118cb8fd" containerID="3afb2ee9befa5aaf5c29206aff62af641a43afa05c3210a05591a2219014eaac" exitCode=0 Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.857663 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" event={"ID":"ff25c029-7927-45c7-bc00-bc1c118cb8fd","Type":"ContainerDied","Data":"3afb2ee9befa5aaf5c29206aff62af641a43afa05c3210a05591a2219014eaac"} Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.866352 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c52xm" event={"ID":"406c250c-ea6b-4820-b38b-6c4d5c0a16de","Type":"ContainerStarted","Data":"2da22b4c8e4a4df292b96bf0ae6ba91d3dee2c12e4bcc3712d6708b54bd2fafa"} Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.867473 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-7hclw" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.868088 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.881934 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8h9mn" Oct 02 09:58:27 crc kubenswrapper[4771]: W1002 09:58:27.897526 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode87072ca_e829_43db_81d0_80df8e4339a7.slice/crio-cb0029d7c722aadae93cb5da316bb7df9339019d0691e3cdfb704f5cec455ed3 WatchSource:0}: Error finding container cb0029d7c722aadae93cb5da316bb7df9339019d0691e3cdfb704f5cec455ed3: Status 404 returned error can't find the container with id cb0029d7c722aadae93cb5da316bb7df9339019d0691e3cdfb704f5cec455ed3 Oct 02 09:58:27 crc kubenswrapper[4771]: I1002 09:58:27.947253 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.092850 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-7hclw"] Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.101146 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-7hclw"] Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.138029 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8h9mn"] Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.160989 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8h9mn"] Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.209948 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 02 09:58:28 crc kubenswrapper[4771]: E1002 09:58:28.212630 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0091b0aa-e43f-476c-8383-c3d279dc11d2" containerName="init" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.212670 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="0091b0aa-e43f-476c-8383-c3d279dc11d2" containerName="init" Oct 02 09:58:28 crc kubenswrapper[4771]: E1002 09:58:28.212744 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4265ff2d-9844-4d4f-922a-efe637c14e8d" containerName="init" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.212750 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4265ff2d-9844-4d4f-922a-efe637c14e8d" containerName="init" Oct 02 09:58:28 crc kubenswrapper[4771]: E1002 09:58:28.212783 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4265ff2d-9844-4d4f-922a-efe637c14e8d" containerName="dnsmasq-dns" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.212790 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4265ff2d-9844-4d4f-922a-efe637c14e8d" containerName="dnsmasq-dns" Oct 02 09:58:28 crc kubenswrapper[4771]: E1002 09:58:28.212848 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0091b0aa-e43f-476c-8383-c3d279dc11d2" containerName="dnsmasq-dns" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.212855 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="0091b0aa-e43f-476c-8383-c3d279dc11d2" containerName="dnsmasq-dns" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.236461 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="0091b0aa-e43f-476c-8383-c3d279dc11d2" containerName="dnsmasq-dns" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.236562 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="4265ff2d-9844-4d4f-922a-efe637c14e8d" containerName="dnsmasq-dns" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.238507 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.238999 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.250037 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-vgwdw" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.255027 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.255428 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.255449 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.429677 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98a17c5b-f1e3-476a-9728-5441e1182f50-scripts\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.429743 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a17c5b-f1e3-476a-9728-5441e1182f50-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.429808 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/98a17c5b-f1e3-476a-9728-5441e1182f50-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.429830 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/98a17c5b-f1e3-476a-9728-5441e1182f50-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.430086 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98a17c5b-f1e3-476a-9728-5441e1182f50-config\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.430306 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/98a17c5b-f1e3-476a-9728-5441e1182f50-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.430672 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j5zt\" (UniqueName: \"kubernetes.io/projected/98a17c5b-f1e3-476a-9728-5441e1182f50-kube-api-access-4j5zt\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.448260 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-qrfml"] Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.449628 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-qrfml" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.463580 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-qrfml"] Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.537163 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j5zt\" (UniqueName: \"kubernetes.io/projected/98a17c5b-f1e3-476a-9728-5441e1182f50-kube-api-access-4j5zt\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.538610 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98a17c5b-f1e3-476a-9728-5441e1182f50-scripts\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.538798 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a17c5b-f1e3-476a-9728-5441e1182f50-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.539037 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/98a17c5b-f1e3-476a-9728-5441e1182f50-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.539163 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/98a17c5b-f1e3-476a-9728-5441e1182f50-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.539330 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98a17c5b-f1e3-476a-9728-5441e1182f50-config\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.539498 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/98a17c5b-f1e3-476a-9728-5441e1182f50-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.539564 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98a17c5b-f1e3-476a-9728-5441e1182f50-scripts\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.540525 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/98a17c5b-f1e3-476a-9728-5441e1182f50-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.540835 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98a17c5b-f1e3-476a-9728-5441e1182f50-config\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.549376 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/98a17c5b-f1e3-476a-9728-5441e1182f50-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.553684 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/98a17c5b-f1e3-476a-9728-5441e1182f50-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.554272 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a17c5b-f1e3-476a-9728-5441e1182f50-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.559287 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j5zt\" (UniqueName: \"kubernetes.io/projected/98a17c5b-f1e3-476a-9728-5441e1182f50-kube-api-access-4j5zt\") pod \"ovn-northd-0\" (UID: \"98a17c5b-f1e3-476a-9728-5441e1182f50\") " pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.641515 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.642265 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjvkk\" (UniqueName: \"kubernetes.io/projected/45eac090-969e-419f-880b-fea01daa6eae-kube-api-access-fjvkk\") pod \"mysqld-exporter-openstack-db-create-qrfml\" (UID: \"45eac090-969e-419f-880b-fea01daa6eae\") " pod="openstack/mysqld-exporter-openstack-db-create-qrfml" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.745049 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjvkk\" (UniqueName: \"kubernetes.io/projected/45eac090-969e-419f-880b-fea01daa6eae-kube-api-access-fjvkk\") pod \"mysqld-exporter-openstack-db-create-qrfml\" (UID: \"45eac090-969e-419f-880b-fea01daa6eae\") " pod="openstack/mysqld-exporter-openstack-db-create-qrfml" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.765707 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjvkk\" (UniqueName: \"kubernetes.io/projected/45eac090-969e-419f-880b-fea01daa6eae-kube-api-access-fjvkk\") pod \"mysqld-exporter-openstack-db-create-qrfml\" (UID: \"45eac090-969e-419f-880b-fea01daa6eae\") " pod="openstack/mysqld-exporter-openstack-db-create-qrfml" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.778348 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-qrfml" Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.901860 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" event={"ID":"ff25c029-7927-45c7-bc00-bc1c118cb8fd","Type":"ContainerStarted","Data":"67d5b132f9f87c6ba20ae5f374f0e9485f95857b88963628860ece04ad218167"} Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.904768 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c52xm" event={"ID":"406c250c-ea6b-4820-b38b-6c4d5c0a16de","Type":"ContainerStarted","Data":"6f70a7be017545691f6e5e7c71cd0b392be013451a5736b10c5ee463ed461a1c"} Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.907736 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" event={"ID":"7eb14f25-cd51-4f54-ac31-3eb9cf51e152","Type":"ContainerStarted","Data":"efa3ef697839c4982a1936f5891ff9ca59a9e10a47012a890cba3c5167d8ec50"} Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.909513 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-s4xgr" event={"ID":"e87072ca-e829-43db-81d0-80df8e4339a7","Type":"ContainerStarted","Data":"00e777980e1c1d12746e7f51d360d96a6dee29e20ae0f960b8b35464ff777036"} Oct 02 09:58:28 crc kubenswrapper[4771]: I1002 09:58:28.909570 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-s4xgr" event={"ID":"e87072ca-e829-43db-81d0-80df8e4339a7","Type":"ContainerStarted","Data":"cb0029d7c722aadae93cb5da316bb7df9339019d0691e3cdfb704f5cec455ed3"} Oct 02 09:58:29 crc kubenswrapper[4771]: I1002 09:58:29.139095 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-x9dc5" podStartSLOduration=4.139067416 podStartE2EDuration="4.139067416s" podCreationTimestamp="2025-10-02 09:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:58:28.954198179 +0000 UTC m=+1296.601883256" watchObservedRunningTime="2025-10-02 09:58:29.139067416 +0000 UTC m=+1296.786752483" Oct 02 09:58:29 crc kubenswrapper[4771]: I1002 09:58:29.141088 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 09:58:29 crc kubenswrapper[4771]: W1002 09:58:29.190769 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98a17c5b_f1e3_476a_9728_5441e1182f50.slice/crio-c9755d14e6253de8c85d5761e1a49f8b4852cb95095d526be9a1a677c03cd0ee WatchSource:0}: Error finding container c9755d14e6253de8c85d5761e1a49f8b4852cb95095d526be9a1a677c03cd0ee: Status 404 returned error can't find the container with id c9755d14e6253de8c85d5761e1a49f8b4852cb95095d526be9a1a677c03cd0ee Oct 02 09:58:29 crc kubenswrapper[4771]: I1002 09:58:29.303823 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-qrfml"] Oct 02 09:58:29 crc kubenswrapper[4771]: W1002 09:58:29.316440 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45eac090_969e_419f_880b_fea01daa6eae.slice/crio-4ff51381bbd8034d2e402538ebf6144e5b81d3db38ad2f77ddfde332816d3746 WatchSource:0}: Error finding container 4ff51381bbd8034d2e402538ebf6144e5b81d3db38ad2f77ddfde332816d3746: Status 404 returned error can't find the container with id 4ff51381bbd8034d2e402538ebf6144e5b81d3db38ad2f77ddfde332816d3746 Oct 02 09:58:29 crc kubenswrapper[4771]: I1002 09:58:29.702340 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0091b0aa-e43f-476c-8383-c3d279dc11d2" path="/var/lib/kubelet/pods/0091b0aa-e43f-476c-8383-c3d279dc11d2/volumes" Oct 02 09:58:29 crc kubenswrapper[4771]: I1002 09:58:29.703848 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4265ff2d-9844-4d4f-922a-efe637c14e8d" path="/var/lib/kubelet/pods/4265ff2d-9844-4d4f-922a-efe637c14e8d/volumes" Oct 02 09:58:29 crc kubenswrapper[4771]: E1002 09:58:29.836755 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod406c250c_ea6b_4820_b38b_6c4d5c0a16de.slice/crio-conmon-6f70a7be017545691f6e5e7c71cd0b392be013451a5736b10c5ee463ed461a1c.scope\": RecentStats: unable to find data in memory cache]" Oct 02 09:58:29 crc kubenswrapper[4771]: I1002 09:58:29.926049 4771 generic.go:334] "Generic (PLEG): container finished" podID="406c250c-ea6b-4820-b38b-6c4d5c0a16de" containerID="6f70a7be017545691f6e5e7c71cd0b392be013451a5736b10c5ee463ed461a1c" exitCode=0 Oct 02 09:58:29 crc kubenswrapper[4771]: I1002 09:58:29.926204 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c52xm" event={"ID":"406c250c-ea6b-4820-b38b-6c4d5c0a16de","Type":"ContainerDied","Data":"6f70a7be017545691f6e5e7c71cd0b392be013451a5736b10c5ee463ed461a1c"} Oct 02 09:58:29 crc kubenswrapper[4771]: I1002 09:58:29.928670 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-qrfml" event={"ID":"45eac090-969e-419f-880b-fea01daa6eae","Type":"ContainerStarted","Data":"4e955b1a9de8071d2de6ea162c7d0e45ab77916019212df4ab3c9eb1b0c2d48c"} Oct 02 09:58:29 crc kubenswrapper[4771]: I1002 09:58:29.928716 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-qrfml" event={"ID":"45eac090-969e-419f-880b-fea01daa6eae","Type":"ContainerStarted","Data":"4ff51381bbd8034d2e402538ebf6144e5b81d3db38ad2f77ddfde332816d3746"} Oct 02 09:58:29 crc kubenswrapper[4771]: I1002 09:58:29.930502 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"98a17c5b-f1e3-476a-9728-5441e1182f50","Type":"ContainerStarted","Data":"c9755d14e6253de8c85d5761e1a49f8b4852cb95095d526be9a1a677c03cd0ee"} Oct 02 09:58:29 crc kubenswrapper[4771]: I1002 09:58:29.973856 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" podStartSLOduration=4.973839912 podStartE2EDuration="4.973839912s" podCreationTimestamp="2025-10-02 09:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:58:29.969794979 +0000 UTC m=+1297.617480046" watchObservedRunningTime="2025-10-02 09:58:29.973839912 +0000 UTC m=+1297.621524979" Oct 02 09:58:29 crc kubenswrapper[4771]: I1002 09:58:29.991833 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-s4xgr" podStartSLOduration=3.9918050899999997 podStartE2EDuration="3.99180509s" podCreationTimestamp="2025-10-02 09:58:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:58:29.985232413 +0000 UTC m=+1297.632917480" watchObservedRunningTime="2025-10-02 09:58:29.99180509 +0000 UTC m=+1297.639490157" Oct 02 09:58:30 crc kubenswrapper[4771]: I1002 09:58:30.014544 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" podStartSLOduration=5.014523021 podStartE2EDuration="5.014523021s" podCreationTimestamp="2025-10-02 09:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:58:30.004635919 +0000 UTC m=+1297.652320976" watchObservedRunningTime="2025-10-02 09:58:30.014523021 +0000 UTC m=+1297.662208088" Oct 02 09:58:30 crc kubenswrapper[4771]: I1002 09:58:30.022366 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-openstack-db-create-qrfml" podStartSLOduration=2.022345711 podStartE2EDuration="2.022345711s" podCreationTimestamp="2025-10-02 09:58:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:58:30.020807572 +0000 UTC m=+1297.668492649" watchObservedRunningTime="2025-10-02 09:58:30.022345711 +0000 UTC m=+1297.670030778" Oct 02 09:58:30 crc kubenswrapper[4771]: I1002 09:58:30.437027 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:30 crc kubenswrapper[4771]: I1002 09:58:30.631493 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:30 crc kubenswrapper[4771]: I1002 09:58:30.942617 4771 generic.go:334] "Generic (PLEG): container finished" podID="45eac090-969e-419f-880b-fea01daa6eae" containerID="4e955b1a9de8071d2de6ea162c7d0e45ab77916019212df4ab3c9eb1b0c2d48c" exitCode=0 Oct 02 09:58:30 crc kubenswrapper[4771]: I1002 09:58:30.942753 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-qrfml" event={"ID":"45eac090-969e-419f-880b-fea01daa6eae","Type":"ContainerDied","Data":"4e955b1a9de8071d2de6ea162c7d0e45ab77916019212df4ab3c9eb1b0c2d48c"} Oct 02 09:58:30 crc kubenswrapper[4771]: I1002 09:58:30.944821 4771 generic.go:334] "Generic (PLEG): container finished" podID="e87072ca-e829-43db-81d0-80df8e4339a7" containerID="00e777980e1c1d12746e7f51d360d96a6dee29e20ae0f960b8b35464ff777036" exitCode=0 Oct 02 09:58:30 crc kubenswrapper[4771]: I1002 09:58:30.944857 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-s4xgr" event={"ID":"e87072ca-e829-43db-81d0-80df8e4339a7","Type":"ContainerDied","Data":"00e777980e1c1d12746e7f51d360d96a6dee29e20ae0f960b8b35464ff777036"} Oct 02 09:58:31 crc kubenswrapper[4771]: I1002 09:58:31.755384 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 02 09:58:31 crc kubenswrapper[4771]: I1002 09:58:31.810163 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c52xm" Oct 02 09:58:31 crc kubenswrapper[4771]: I1002 09:58:31.921678 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xc4j\" (UniqueName: \"kubernetes.io/projected/406c250c-ea6b-4820-b38b-6c4d5c0a16de-kube-api-access-7xc4j\") pod \"406c250c-ea6b-4820-b38b-6c4d5c0a16de\" (UID: \"406c250c-ea6b-4820-b38b-6c4d5c0a16de\") " Oct 02 09:58:31 crc kubenswrapper[4771]: I1002 09:58:31.933495 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/406c250c-ea6b-4820-b38b-6c4d5c0a16de-kube-api-access-7xc4j" (OuterVolumeSpecName: "kube-api-access-7xc4j") pod "406c250c-ea6b-4820-b38b-6c4d5c0a16de" (UID: "406c250c-ea6b-4820-b38b-6c4d5c0a16de"). InnerVolumeSpecName "kube-api-access-7xc4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:31 crc kubenswrapper[4771]: I1002 09:58:31.999522 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c52xm" Oct 02 09:58:32 crc kubenswrapper[4771]: I1002 09:58:32.000484 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c52xm" event={"ID":"406c250c-ea6b-4820-b38b-6c4d5c0a16de","Type":"ContainerDied","Data":"2da22b4c8e4a4df292b96bf0ae6ba91d3dee2c12e4bcc3712d6708b54bd2fafa"} Oct 02 09:58:32 crc kubenswrapper[4771]: I1002 09:58:32.000549 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2da22b4c8e4a4df292b96bf0ae6ba91d3dee2c12e4bcc3712d6708b54bd2fafa" Oct 02 09:58:32 crc kubenswrapper[4771]: I1002 09:58:32.001740 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-xcv4v"] Oct 02 09:58:32 crc kubenswrapper[4771]: E1002 09:58:32.002309 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="406c250c-ea6b-4820-b38b-6c4d5c0a16de" containerName="mariadb-database-create" Oct 02 09:58:32 crc kubenswrapper[4771]: I1002 09:58:32.002328 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="406c250c-ea6b-4820-b38b-6c4d5c0a16de" containerName="mariadb-database-create" Oct 02 09:58:32 crc kubenswrapper[4771]: I1002 09:58:32.002595 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="406c250c-ea6b-4820-b38b-6c4d5c0a16de" containerName="mariadb-database-create" Oct 02 09:58:32 crc kubenswrapper[4771]: I1002 09:58:32.003489 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xcv4v" Oct 02 09:58:32 crc kubenswrapper[4771]: I1002 09:58:32.025275 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xc4j\" (UniqueName: \"kubernetes.io/projected/406c250c-ea6b-4820-b38b-6c4d5c0a16de-kube-api-access-7xc4j\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:32 crc kubenswrapper[4771]: I1002 09:58:32.028148 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xcv4v"] Oct 02 09:58:32 crc kubenswrapper[4771]: I1002 09:58:32.127260 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvjb5\" (UniqueName: \"kubernetes.io/projected/b0b9f6fb-1605-49b6-865d-3945530f7848-kube-api-access-qvjb5\") pod \"glance-db-create-xcv4v\" (UID: \"b0b9f6fb-1605-49b6-865d-3945530f7848\") " pod="openstack/glance-db-create-xcv4v" Oct 02 09:58:32 crc kubenswrapper[4771]: I1002 09:58:32.229873 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvjb5\" (UniqueName: \"kubernetes.io/projected/b0b9f6fb-1605-49b6-865d-3945530f7848-kube-api-access-qvjb5\") pod \"glance-db-create-xcv4v\" (UID: \"b0b9f6fb-1605-49b6-865d-3945530f7848\") " pod="openstack/glance-db-create-xcv4v" Oct 02 09:58:32 crc kubenswrapper[4771]: I1002 09:58:32.262577 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvjb5\" (UniqueName: \"kubernetes.io/projected/b0b9f6fb-1605-49b6-865d-3945530f7848-kube-api-access-qvjb5\") pod \"glance-db-create-xcv4v\" (UID: \"b0b9f6fb-1605-49b6-865d-3945530f7848\") " pod="openstack/glance-db-create-xcv4v" Oct 02 09:58:33 crc kubenswrapper[4771]: E1002 09:58:32.405972 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98a17c5b_f1e3_476a_9728_5441e1182f50.slice/crio-49a4375c76c729f1c755d06b6f78a0f4acfea7f278e82ccea27bf1110b1c01d0.scope\": RecentStats: unable to find data in memory cache]" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:32.440357 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xcv4v" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:32.623178 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s4xgr" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:32.631219 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-qrfml" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:32.768238 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr7jn\" (UniqueName: \"kubernetes.io/projected/e87072ca-e829-43db-81d0-80df8e4339a7-kube-api-access-pr7jn\") pod \"e87072ca-e829-43db-81d0-80df8e4339a7\" (UID: \"e87072ca-e829-43db-81d0-80df8e4339a7\") " Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:32.769115 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjvkk\" (UniqueName: \"kubernetes.io/projected/45eac090-969e-419f-880b-fea01daa6eae-kube-api-access-fjvkk\") pod \"45eac090-969e-419f-880b-fea01daa6eae\" (UID: \"45eac090-969e-419f-880b-fea01daa6eae\") " Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:32.778333 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45eac090-969e-419f-880b-fea01daa6eae-kube-api-access-fjvkk" (OuterVolumeSpecName: "kube-api-access-fjvkk") pod "45eac090-969e-419f-880b-fea01daa6eae" (UID: "45eac090-969e-419f-880b-fea01daa6eae"). InnerVolumeSpecName "kube-api-access-fjvkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:32.778578 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e87072ca-e829-43db-81d0-80df8e4339a7-kube-api-access-pr7jn" (OuterVolumeSpecName: "kube-api-access-pr7jn") pod "e87072ca-e829-43db-81d0-80df8e4339a7" (UID: "e87072ca-e829-43db-81d0-80df8e4339a7"). InnerVolumeSpecName "kube-api-access-pr7jn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:32.875015 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjvkk\" (UniqueName: \"kubernetes.io/projected/45eac090-969e-419f-880b-fea01daa6eae-kube-api-access-fjvkk\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:32.875054 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr7jn\" (UniqueName: \"kubernetes.io/projected/e87072ca-e829-43db-81d0-80df8e4339a7-kube-api-access-pr7jn\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:33.011384 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s4xgr" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:33.011389 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-s4xgr" event={"ID":"e87072ca-e829-43db-81d0-80df8e4339a7","Type":"ContainerDied","Data":"cb0029d7c722aadae93cb5da316bb7df9339019d0691e3cdfb704f5cec455ed3"} Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:33.011557 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb0029d7c722aadae93cb5da316bb7df9339019d0691e3cdfb704f5cec455ed3" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:33.012854 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-qrfml" event={"ID":"45eac090-969e-419f-880b-fea01daa6eae","Type":"ContainerDied","Data":"4ff51381bbd8034d2e402538ebf6144e5b81d3db38ad2f77ddfde332816d3746"} Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:33.012865 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-qrfml" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:33.012892 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ff51381bbd8034d2e402538ebf6144e5b81d3db38ad2f77ddfde332816d3746" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:33.018625 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"98a17c5b-f1e3-476a-9728-5441e1182f50","Type":"ContainerStarted","Data":"49a4375c76c729f1c755d06b6f78a0f4acfea7f278e82ccea27bf1110b1c01d0"} Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:33.018669 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"98a17c5b-f1e3-476a-9728-5441e1182f50","Type":"ContainerStarted","Data":"02aa4ef21fe7d741b962fa2a8fcd45398d79a87dd1434154e04edb9b557fe5b0"} Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:33.019920 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:33.511449 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.873335003 podStartE2EDuration="5.511425021s" podCreationTimestamp="2025-10-02 09:58:28 +0000 UTC" firstStartedPulling="2025-10-02 09:58:29.200381984 +0000 UTC m=+1296.848067051" lastFinishedPulling="2025-10-02 09:58:31.838472002 +0000 UTC m=+1299.486157069" observedRunningTime="2025-10-02 09:58:33.050577197 +0000 UTC m=+1300.698262264" watchObservedRunningTime="2025-10-02 09:58:33.511425021 +0000 UTC m=+1301.159110088" Oct 02 09:58:33 crc kubenswrapper[4771]: I1002 09:58:33.513935 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xcv4v"] Oct 02 09:58:34 crc kubenswrapper[4771]: I1002 09:58:34.032501 4771 generic.go:334] "Generic (PLEG): container finished" podID="b0b9f6fb-1605-49b6-865d-3945530f7848" containerID="4d78210951b47065d8b907ebe3e9c60ba32b31de8da47258db41376df9122714" exitCode=0 Oct 02 09:58:34 crc kubenswrapper[4771]: I1002 09:58:34.032561 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xcv4v" event={"ID":"b0b9f6fb-1605-49b6-865d-3945530f7848","Type":"ContainerDied","Data":"4d78210951b47065d8b907ebe3e9c60ba32b31de8da47258db41376df9122714"} Oct 02 09:58:34 crc kubenswrapper[4771]: I1002 09:58:34.032826 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xcv4v" event={"ID":"b0b9f6fb-1605-49b6-865d-3945530f7848","Type":"ContainerStarted","Data":"bd95395774bdbec45d7d064c436341aaedb36db18c1c22bbe333da9499408779"} Oct 02 09:58:35 crc kubenswrapper[4771]: I1002 09:58:35.433171 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:35 crc kubenswrapper[4771]: I1002 09:58:35.632998 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:35 crc kubenswrapper[4771]: I1002 09:58:35.697000 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xcv4v" Oct 02 09:58:35 crc kubenswrapper[4771]: I1002 09:58:35.723062 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-vcdhf"] Oct 02 09:58:35 crc kubenswrapper[4771]: I1002 09:58:35.747059 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvjb5\" (UniqueName: \"kubernetes.io/projected/b0b9f6fb-1605-49b6-865d-3945530f7848-kube-api-access-qvjb5\") pod \"b0b9f6fb-1605-49b6-865d-3945530f7848\" (UID: \"b0b9f6fb-1605-49b6-865d-3945530f7848\") " Oct 02 09:58:35 crc kubenswrapper[4771]: I1002 09:58:35.761056 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0b9f6fb-1605-49b6-865d-3945530f7848-kube-api-access-qvjb5" (OuterVolumeSpecName: "kube-api-access-qvjb5") pod "b0b9f6fb-1605-49b6-865d-3945530f7848" (UID: "b0b9f6fb-1605-49b6-865d-3945530f7848"). InnerVolumeSpecName "kube-api-access-qvjb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:35 crc kubenswrapper[4771]: I1002 09:58:35.851491 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvjb5\" (UniqueName: \"kubernetes.io/projected/b0b9f6fb-1605-49b6-865d-3945530f7848-kube-api-access-qvjb5\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.058480 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" podUID="ff25c029-7927-45c7-bc00-bc1c118cb8fd" containerName="dnsmasq-dns" containerID="cri-o://67d5b132f9f87c6ba20ae5f374f0e9485f95857b88963628860ece04ad218167" gracePeriod=10 Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.058705 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xcv4v" Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.062049 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xcv4v" event={"ID":"b0b9f6fb-1605-49b6-865d-3945530f7848","Type":"ContainerDied","Data":"bd95395774bdbec45d7d064c436341aaedb36db18c1c22bbe333da9499408779"} Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.062223 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd95395774bdbec45d7d064c436341aaedb36db18c1c22bbe333da9499408779" Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.793750 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.873875 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxkr2\" (UniqueName: \"kubernetes.io/projected/ff25c029-7927-45c7-bc00-bc1c118cb8fd-kube-api-access-pxkr2\") pod \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.874186 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-dns-svc\") pod \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.874292 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-config\") pod \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.874321 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-ovsdbserver-sb\") pod \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\" (UID: \"ff25c029-7927-45c7-bc00-bc1c118cb8fd\") " Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.881921 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff25c029-7927-45c7-bc00-bc1c118cb8fd-kube-api-access-pxkr2" (OuterVolumeSpecName: "kube-api-access-pxkr2") pod "ff25c029-7927-45c7-bc00-bc1c118cb8fd" (UID: "ff25c029-7927-45c7-bc00-bc1c118cb8fd"). InnerVolumeSpecName "kube-api-access-pxkr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.938971 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ff25c029-7927-45c7-bc00-bc1c118cb8fd" (UID: "ff25c029-7927-45c7-bc00-bc1c118cb8fd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.939390 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ff25c029-7927-45c7-bc00-bc1c118cb8fd" (UID: "ff25c029-7927-45c7-bc00-bc1c118cb8fd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.944967 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-config" (OuterVolumeSpecName: "config") pod "ff25c029-7927-45c7-bc00-bc1c118cb8fd" (UID: "ff25c029-7927-45c7-bc00-bc1c118cb8fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.977219 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.977271 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.977284 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff25c029-7927-45c7-bc00-bc1c118cb8fd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:36 crc kubenswrapper[4771]: I1002 09:58:36.977305 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxkr2\" (UniqueName: \"kubernetes.io/projected/ff25c029-7927-45c7-bc00-bc1c118cb8fd-kube-api-access-pxkr2\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.071268 4771 generic.go:334] "Generic (PLEG): container finished" podID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerID="6fd70e48c855a5f9bf81cbf37919a8679ecbde45a91dccd0e2a6fc1a67916fa6" exitCode=0 Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.071344 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"95cd3d5a-1000-4212-b4d9-5d0da549c7f3","Type":"ContainerDied","Data":"6fd70e48c855a5f9bf81cbf37919a8679ecbde45a91dccd0e2a6fc1a67916fa6"} Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.075430 4771 generic.go:334] "Generic (PLEG): container finished" podID="ff25c029-7927-45c7-bc00-bc1c118cb8fd" containerID="67d5b132f9f87c6ba20ae5f374f0e9485f95857b88963628860ece04ad218167" exitCode=0 Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.075481 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" event={"ID":"ff25c029-7927-45c7-bc00-bc1c118cb8fd","Type":"ContainerDied","Data":"67d5b132f9f87c6ba20ae5f374f0e9485f95857b88963628860ece04ad218167"} Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.075525 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" event={"ID":"ff25c029-7927-45c7-bc00-bc1c118cb8fd","Type":"ContainerDied","Data":"ecafcbf06a969847abe1fc9ee58557183461028a15dc95123269e76d6ef3da3c"} Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.075538 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-vcdhf" Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.075549 4771 scope.go:117] "RemoveContainer" containerID="67d5b132f9f87c6ba20ae5f374f0e9485f95857b88963628860ece04ad218167" Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.128401 4771 scope.go:117] "RemoveContainer" containerID="3afb2ee9befa5aaf5c29206aff62af641a43afa05c3210a05591a2219014eaac" Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.173925 4771 scope.go:117] "RemoveContainer" containerID="67d5b132f9f87c6ba20ae5f374f0e9485f95857b88963628860ece04ad218167" Oct 02 09:58:37 crc kubenswrapper[4771]: E1002 09:58:37.174470 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67d5b132f9f87c6ba20ae5f374f0e9485f95857b88963628860ece04ad218167\": container with ID starting with 67d5b132f9f87c6ba20ae5f374f0e9485f95857b88963628860ece04ad218167 not found: ID does not exist" containerID="67d5b132f9f87c6ba20ae5f374f0e9485f95857b88963628860ece04ad218167" Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.174518 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67d5b132f9f87c6ba20ae5f374f0e9485f95857b88963628860ece04ad218167"} err="failed to get container status \"67d5b132f9f87c6ba20ae5f374f0e9485f95857b88963628860ece04ad218167\": rpc error: code = NotFound desc = could not find container \"67d5b132f9f87c6ba20ae5f374f0e9485f95857b88963628860ece04ad218167\": container with ID starting with 67d5b132f9f87c6ba20ae5f374f0e9485f95857b88963628860ece04ad218167 not found: ID does not exist" Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.174544 4771 scope.go:117] "RemoveContainer" containerID="3afb2ee9befa5aaf5c29206aff62af641a43afa05c3210a05591a2219014eaac" Oct 02 09:58:37 crc kubenswrapper[4771]: E1002 09:58:37.175097 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3afb2ee9befa5aaf5c29206aff62af641a43afa05c3210a05591a2219014eaac\": container with ID starting with 3afb2ee9befa5aaf5c29206aff62af641a43afa05c3210a05591a2219014eaac not found: ID does not exist" containerID="3afb2ee9befa5aaf5c29206aff62af641a43afa05c3210a05591a2219014eaac" Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.175159 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afb2ee9befa5aaf5c29206aff62af641a43afa05c3210a05591a2219014eaac"} err="failed to get container status \"3afb2ee9befa5aaf5c29206aff62af641a43afa05c3210a05591a2219014eaac\": rpc error: code = NotFound desc = could not find container \"3afb2ee9befa5aaf5c29206aff62af641a43afa05c3210a05591a2219014eaac\": container with ID starting with 3afb2ee9befa5aaf5c29206aff62af641a43afa05c3210a05591a2219014eaac not found: ID does not exist" Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.184338 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-vcdhf"] Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.195880 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-vcdhf"] Oct 02 09:58:37 crc kubenswrapper[4771]: I1002 09:58:37.696687 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff25c029-7927-45c7-bc00-bc1c118cb8fd" path="/var/lib/kubelet/pods/ff25c029-7927-45c7-bc00-bc1c118cb8fd/volumes" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.002311 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-5jf58" podUID="3388e290-ed26-4869-b6ce-c93670e7d9cf" containerName="ovn-controller" probeResult="failure" output=< Oct 02 09:58:38 crc kubenswrapper[4771]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 09:58:38 crc kubenswrapper[4771]: > Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.376362 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.385903 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-9qhq9" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.661831 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5jf58-config-4rvzp"] Oct 02 09:58:38 crc kubenswrapper[4771]: E1002 09:58:38.663033 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0b9f6fb-1605-49b6-865d-3945530f7848" containerName="mariadb-database-create" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.663063 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0b9f6fb-1605-49b6-865d-3945530f7848" containerName="mariadb-database-create" Oct 02 09:58:38 crc kubenswrapper[4771]: E1002 09:58:38.663097 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45eac090-969e-419f-880b-fea01daa6eae" containerName="mariadb-database-create" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.663110 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="45eac090-969e-419f-880b-fea01daa6eae" containerName="mariadb-database-create" Oct 02 09:58:38 crc kubenswrapper[4771]: E1002 09:58:38.663144 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff25c029-7927-45c7-bc00-bc1c118cb8fd" containerName="init" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.663152 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff25c029-7927-45c7-bc00-bc1c118cb8fd" containerName="init" Oct 02 09:58:38 crc kubenswrapper[4771]: E1002 09:58:38.663173 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e87072ca-e829-43db-81d0-80df8e4339a7" containerName="mariadb-database-create" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.663179 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e87072ca-e829-43db-81d0-80df8e4339a7" containerName="mariadb-database-create" Oct 02 09:58:38 crc kubenswrapper[4771]: E1002 09:58:38.663195 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff25c029-7927-45c7-bc00-bc1c118cb8fd" containerName="dnsmasq-dns" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.663204 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff25c029-7927-45c7-bc00-bc1c118cb8fd" containerName="dnsmasq-dns" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.663512 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0b9f6fb-1605-49b6-865d-3945530f7848" containerName="mariadb-database-create" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.663551 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="e87072ca-e829-43db-81d0-80df8e4339a7" containerName="mariadb-database-create" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.663586 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="45eac090-969e-419f-880b-fea01daa6eae" containerName="mariadb-database-create" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.663600 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff25c029-7927-45c7-bc00-bc1c118cb8fd" containerName="dnsmasq-dns" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.664836 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.671232 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.685686 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5jf58-config-4rvzp"] Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.713469 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-run\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.715471 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-log-ovn\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.715589 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/514ca08b-64a5-457d-ac4f-e763ab9ad718-scripts\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.727967 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/514ca08b-64a5-457d-ac4f-e763ab9ad718-additional-scripts\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.728064 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-run-ovn\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.728206 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss7pl\" (UniqueName: \"kubernetes.io/projected/514ca08b-64a5-457d-ac4f-e763ab9ad718-kube-api-access-ss7pl\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.760002 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-bjg2r"] Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.762597 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.792056 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-bjg2r"] Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.822808 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-5bfb-account-create-sf2wc"] Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.824692 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-5bfb-account-create-sf2wc" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.829768 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.840738 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-config\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.840815 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-log-ovn\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.840851 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.840893 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/514ca08b-64a5-457d-ac4f-e763ab9ad718-scripts\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.840977 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2lw8\" (UniqueName: \"kubernetes.io/projected/e786364e-fd43-43ed-a371-b87cc5a1e59a-kube-api-access-j2lw8\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.841028 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/514ca08b-64a5-457d-ac4f-e763ab9ad718-additional-scripts\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.841053 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-run-ovn\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.841499 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.841581 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-run-ovn\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.841623 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss7pl\" (UniqueName: \"kubernetes.io/projected/514ca08b-64a5-457d-ac4f-e763ab9ad718-kube-api-access-ss7pl\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.842201 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/514ca08b-64a5-457d-ac4f-e763ab9ad718-additional-scripts\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.842991 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-dns-svc\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.843159 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-run\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.846618 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-run\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.846726 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-log-ovn\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.850344 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-5bfb-account-create-sf2wc"] Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.855191 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/514ca08b-64a5-457d-ac4f-e763ab9ad718-scripts\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.873078 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss7pl\" (UniqueName: \"kubernetes.io/projected/514ca08b-64a5-457d-ac4f-e763ab9ad718-kube-api-access-ss7pl\") pod \"ovn-controller-5jf58-config-4rvzp\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.948092 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.948252 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wg6m\" (UniqueName: \"kubernetes.io/projected/e8507c3c-8ba8-47e3-9529-503679e4ffa7-kube-api-access-9wg6m\") pod \"mysqld-exporter-5bfb-account-create-sf2wc\" (UID: \"e8507c3c-8ba8-47e3-9529-503679e4ffa7\") " pod="openstack/mysqld-exporter-5bfb-account-create-sf2wc" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.948298 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2lw8\" (UniqueName: \"kubernetes.io/projected/e786364e-fd43-43ed-a371-b87cc5a1e59a-kube-api-access-j2lw8\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.948372 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.948480 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-dns-svc\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.948574 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-config\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.949530 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-config\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.950192 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.951200 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.951616 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-dns-svc\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:38 crc kubenswrapper[4771]: I1002 09:58:38.978742 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2lw8\" (UniqueName: \"kubernetes.io/projected/e786364e-fd43-43ed-a371-b87cc5a1e59a-kube-api-access-j2lw8\") pod \"dnsmasq-dns-698758b865-bjg2r\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:39 crc kubenswrapper[4771]: I1002 09:58:39.005826 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:39 crc kubenswrapper[4771]: I1002 09:58:39.055353 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wg6m\" (UniqueName: \"kubernetes.io/projected/e8507c3c-8ba8-47e3-9529-503679e4ffa7-kube-api-access-9wg6m\") pod \"mysqld-exporter-5bfb-account-create-sf2wc\" (UID: \"e8507c3c-8ba8-47e3-9529-503679e4ffa7\") " pod="openstack/mysqld-exporter-5bfb-account-create-sf2wc" Oct 02 09:58:39 crc kubenswrapper[4771]: I1002 09:58:39.101433 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:39 crc kubenswrapper[4771]: I1002 09:58:39.109799 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wg6m\" (UniqueName: \"kubernetes.io/projected/e8507c3c-8ba8-47e3-9529-503679e4ffa7-kube-api-access-9wg6m\") pod \"mysqld-exporter-5bfb-account-create-sf2wc\" (UID: \"e8507c3c-8ba8-47e3-9529-503679e4ffa7\") " pod="openstack/mysqld-exporter-5bfb-account-create-sf2wc" Oct 02 09:58:39 crc kubenswrapper[4771]: I1002 09:58:39.150486 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-5bfb-account-create-sf2wc" Oct 02 09:58:39 crc kubenswrapper[4771]: I1002 09:58:39.610950 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5jf58-config-4rvzp"] Oct 02 09:58:39 crc kubenswrapper[4771]: I1002 09:58:39.754584 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-bjg2r"] Oct 02 09:58:39 crc kubenswrapper[4771]: W1002 09:58:39.772005 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode786364e_fd43_43ed_a371_b87cc5a1e59a.slice/crio-c6d8b754cd91ab483faff42b6c009f6abc6b3d35e4821bded38e2112de0153cd WatchSource:0}: Error finding container c6d8b754cd91ab483faff42b6c009f6abc6b3d35e4821bded38e2112de0153cd: Status 404 returned error can't find the container with id c6d8b754cd91ab483faff42b6c009f6abc6b3d35e4821bded38e2112de0153cd Oct 02 09:58:39 crc kubenswrapper[4771]: I1002 09:58:39.926954 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-5bfb-account-create-sf2wc"] Oct 02 09:58:39 crc kubenswrapper[4771]: I1002 09:58:39.987389 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 02 09:58:39 crc kubenswrapper[4771]: I1002 09:58:39.996567 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.001424 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.010946 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.011289 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.011448 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.011614 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-wfn8f" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.102433 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-lock\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.102999 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksdtc\" (UniqueName: \"kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-kube-api-access-ksdtc\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.103036 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-cache\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.103074 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.103167 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.136414 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5jf58-config-4rvzp" event={"ID":"514ca08b-64a5-457d-ac4f-e763ab9ad718","Type":"ContainerStarted","Data":"404f77d21985e3ea26d86aded5b1ee628311bb987be55bed87d685271f5282bb"} Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.146106 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-5bfb-account-create-sf2wc" event={"ID":"e8507c3c-8ba8-47e3-9529-503679e4ffa7","Type":"ContainerStarted","Data":"9c5404d5ea89531d44b1d65193b73946ab5d2fc8b281ad26e1cd7c41af6230ac"} Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.150554 4771 generic.go:334] "Generic (PLEG): container finished" podID="b7dad196-4ebc-43e0-b850-0820f2b072c9" containerID="63aadd758b05567dc1365ba3e674b8ba28ce7f2c61ec9a5fcd81f0d9682d8a7c" exitCode=0 Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.150640 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7dad196-4ebc-43e0-b850-0820f2b072c9","Type":"ContainerDied","Data":"63aadd758b05567dc1365ba3e674b8ba28ce7f2c61ec9a5fcd81f0d9682d8a7c"} Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.156620 4771 generic.go:334] "Generic (PLEG): container finished" podID="ebc075f5-e2ba-4789-a38e-4065fe661ada" containerID="74e0addf34252d1465b8adca2fc82ba42dcd867d6b03aac7aec4c69a2f14976d" exitCode=0 Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.156726 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ebc075f5-e2ba-4789-a38e-4065fe661ada","Type":"ContainerDied","Data":"74e0addf34252d1465b8adca2fc82ba42dcd867d6b03aac7aec4c69a2f14976d"} Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.163342 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-bjg2r" event={"ID":"e786364e-fd43-43ed-a371-b87cc5a1e59a","Type":"ContainerStarted","Data":"c6d8b754cd91ab483faff42b6c009f6abc6b3d35e4821bded38e2112de0153cd"} Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.205961 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-lock\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.206136 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksdtc\" (UniqueName: \"kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-kube-api-access-ksdtc\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.206193 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-cache\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.206248 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.206453 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.208207 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-lock\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: E1002 09:58:40.209489 4771 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 09:58:40 crc kubenswrapper[4771]: E1002 09:58:40.219446 4771 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 09:58:40 crc kubenswrapper[4771]: E1002 09:58:40.219684 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift podName:4c4deacb-7b2f-49d1-8c7e-b87913f027d1 nodeName:}" failed. No retries permitted until 2025-10-02 09:58:40.719643375 +0000 UTC m=+1308.367328442 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift") pod "swift-storage-0" (UID: "4c4deacb-7b2f-49d1-8c7e-b87913f027d1") : configmap "swift-ring-files" not found Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.209629 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.209614 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-cache\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.230501 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksdtc\" (UniqueName: \"kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-kube-api-access-ksdtc\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.308306 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: I1002 09:58:40.723041 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:40 crc kubenswrapper[4771]: E1002 09:58:40.723949 4771 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 09:58:40 crc kubenswrapper[4771]: E1002 09:58:40.723998 4771 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 09:58:40 crc kubenswrapper[4771]: E1002 09:58:40.724082 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift podName:4c4deacb-7b2f-49d1-8c7e-b87913f027d1 nodeName:}" failed. No retries permitted until 2025-10-02 09:58:41.724054734 +0000 UTC m=+1309.371739801 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift") pod "swift-storage-0" (UID: "4c4deacb-7b2f-49d1-8c7e-b87913f027d1") : configmap "swift-ring-files" not found Oct 02 09:58:41 crc kubenswrapper[4771]: I1002 09:58:41.191940 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ebc075f5-e2ba-4789-a38e-4065fe661ada","Type":"ContainerStarted","Data":"82fe7b3800926b8bc7dfa8813efaf7f0dd9c638720dde9673e1bf775cfa6e1dd"} Oct 02 09:58:41 crc kubenswrapper[4771]: I1002 09:58:41.192482 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 09:58:41 crc kubenswrapper[4771]: I1002 09:58:41.195886 4771 generic.go:334] "Generic (PLEG): container finished" podID="e786364e-fd43-43ed-a371-b87cc5a1e59a" containerID="f401f8e208415441627053abbaf61bd8fa3facefcad79e96174e420ca6f40b3c" exitCode=0 Oct 02 09:58:41 crc kubenswrapper[4771]: I1002 09:58:41.195983 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-bjg2r" event={"ID":"e786364e-fd43-43ed-a371-b87cc5a1e59a","Type":"ContainerDied","Data":"f401f8e208415441627053abbaf61bd8fa3facefcad79e96174e420ca6f40b3c"} Oct 02 09:58:41 crc kubenswrapper[4771]: I1002 09:58:41.203591 4771 generic.go:334] "Generic (PLEG): container finished" podID="514ca08b-64a5-457d-ac4f-e763ab9ad718" containerID="97f487f824331b0f6a26a035064204af8d304d5c05bdbf68aa793206ac3d929c" exitCode=0 Oct 02 09:58:41 crc kubenswrapper[4771]: I1002 09:58:41.203682 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5jf58-config-4rvzp" event={"ID":"514ca08b-64a5-457d-ac4f-e763ab9ad718","Type":"ContainerDied","Data":"97f487f824331b0f6a26a035064204af8d304d5c05bdbf68aa793206ac3d929c"} Oct 02 09:58:41 crc kubenswrapper[4771]: I1002 09:58:41.207461 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7dad196-4ebc-43e0-b850-0820f2b072c9","Type":"ContainerStarted","Data":"600abb9d49fe9d1c7f960409891f3dd4c17c8d27d72e596b2ef8946d71c7edff"} Oct 02 09:58:41 crc kubenswrapper[4771]: I1002 09:58:41.207847 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:58:41 crc kubenswrapper[4771]: I1002 09:58:41.236403 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.341412237 podStartE2EDuration="1m19.236376723s" podCreationTimestamp="2025-10-02 09:57:22 +0000 UTC" firstStartedPulling="2025-10-02 09:57:24.704501343 +0000 UTC m=+1232.352186400" lastFinishedPulling="2025-10-02 09:58:05.599465829 +0000 UTC m=+1273.247150886" observedRunningTime="2025-10-02 09:58:41.22450183 +0000 UTC m=+1308.872186907" watchObservedRunningTime="2025-10-02 09:58:41.236376723 +0000 UTC m=+1308.884061790" Oct 02 09:58:41 crc kubenswrapper[4771]: I1002 09:58:41.238305 4771 generic.go:334] "Generic (PLEG): container finished" podID="e8507c3c-8ba8-47e3-9529-503679e4ffa7" containerID="7c2a21faff2c4fe7a008af7edd09446ca1f6719bbd3ae75737b5bf0415b5e8d3" exitCode=0 Oct 02 09:58:41 crc kubenswrapper[4771]: I1002 09:58:41.238381 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-5bfb-account-create-sf2wc" event={"ID":"e8507c3c-8ba8-47e3-9529-503679e4ffa7","Type":"ContainerDied","Data":"7c2a21faff2c4fe7a008af7edd09446ca1f6719bbd3ae75737b5bf0415b5e8d3"} Oct 02 09:58:41 crc kubenswrapper[4771]: I1002 09:58:41.369776 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.817535232 podStartE2EDuration="1m20.369750204s" podCreationTimestamp="2025-10-02 09:57:21 +0000 UTC" firstStartedPulling="2025-10-02 09:57:24.051473005 +0000 UTC m=+1231.699158072" lastFinishedPulling="2025-10-02 09:58:05.603687977 +0000 UTC m=+1273.251373044" observedRunningTime="2025-10-02 09:58:41.352917934 +0000 UTC m=+1309.000603011" watchObservedRunningTime="2025-10-02 09:58:41.369750204 +0000 UTC m=+1309.017435281" Oct 02 09:58:41 crc kubenswrapper[4771]: I1002 09:58:41.753326 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:41 crc kubenswrapper[4771]: E1002 09:58:41.754103 4771 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 09:58:41 crc kubenswrapper[4771]: E1002 09:58:41.754165 4771 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 09:58:41 crc kubenswrapper[4771]: E1002 09:58:41.754246 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift podName:4c4deacb-7b2f-49d1-8c7e-b87913f027d1 nodeName:}" failed. No retries permitted until 2025-10-02 09:58:43.754217015 +0000 UTC m=+1311.401902072 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift") pod "swift-storage-0" (UID: "4c4deacb-7b2f-49d1-8c7e-b87913f027d1") : configmap "swift-ring-files" not found Oct 02 09:58:42 crc kubenswrapper[4771]: I1002 09:58:42.028367 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-fb66-account-create-q85l7"] Oct 02 09:58:42 crc kubenswrapper[4771]: I1002 09:58:42.031648 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fb66-account-create-q85l7" Oct 02 09:58:42 crc kubenswrapper[4771]: I1002 09:58:42.035224 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 02 09:58:42 crc kubenswrapper[4771]: I1002 09:58:42.050328 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fb66-account-create-q85l7"] Oct 02 09:58:42 crc kubenswrapper[4771]: I1002 09:58:42.146526 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:58:42 crc kubenswrapper[4771]: I1002 09:58:42.146602 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:58:42 crc kubenswrapper[4771]: I1002 09:58:42.166379 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjk5k\" (UniqueName: \"kubernetes.io/projected/1ba7c27c-2a56-4f63-9082-80935e39a0e2-kube-api-access-pjk5k\") pod \"glance-fb66-account-create-q85l7\" (UID: \"1ba7c27c-2a56-4f63-9082-80935e39a0e2\") " pod="openstack/glance-fb66-account-create-q85l7" Oct 02 09:58:42 crc kubenswrapper[4771]: I1002 09:58:42.256624 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-bjg2r" event={"ID":"e786364e-fd43-43ed-a371-b87cc5a1e59a","Type":"ContainerStarted","Data":"db107d781bec174e1922a04a1d846e1cb676be4c68153f675441a3c53dc84bdb"} Oct 02 09:58:42 crc kubenswrapper[4771]: I1002 09:58:42.256815 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:42 crc kubenswrapper[4771]: I1002 09:58:42.270675 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjk5k\" (UniqueName: \"kubernetes.io/projected/1ba7c27c-2a56-4f63-9082-80935e39a0e2-kube-api-access-pjk5k\") pod \"glance-fb66-account-create-q85l7\" (UID: \"1ba7c27c-2a56-4f63-9082-80935e39a0e2\") " pod="openstack/glance-fb66-account-create-q85l7" Oct 02 09:58:42 crc kubenswrapper[4771]: I1002 09:58:42.295154 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-bjg2r" podStartSLOduration=4.295086626 podStartE2EDuration="4.295086626s" podCreationTimestamp="2025-10-02 09:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:58:42.284500865 +0000 UTC m=+1309.932185922" watchObservedRunningTime="2025-10-02 09:58:42.295086626 +0000 UTC m=+1309.942771693" Oct 02 09:58:42 crc kubenswrapper[4771]: I1002 09:58:42.317371 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjk5k\" (UniqueName: \"kubernetes.io/projected/1ba7c27c-2a56-4f63-9082-80935e39a0e2-kube-api-access-pjk5k\") pod \"glance-fb66-account-create-q85l7\" (UID: \"1ba7c27c-2a56-4f63-9082-80935e39a0e2\") " pod="openstack/glance-fb66-account-create-q85l7" Oct 02 09:58:42 crc kubenswrapper[4771]: I1002 09:58:42.374179 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fb66-account-create-q85l7" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.058330 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-5jf58" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.237954 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-5bfb-account-create-sf2wc" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.311902 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wg6m\" (UniqueName: \"kubernetes.io/projected/e8507c3c-8ba8-47e3-9529-503679e4ffa7-kube-api-access-9wg6m\") pod \"e8507c3c-8ba8-47e3-9529-503679e4ffa7\" (UID: \"e8507c3c-8ba8-47e3-9529-503679e4ffa7\") " Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.316579 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-5bfb-account-create-sf2wc" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.317067 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-5bfb-account-create-sf2wc" event={"ID":"e8507c3c-8ba8-47e3-9529-503679e4ffa7","Type":"ContainerDied","Data":"9c5404d5ea89531d44b1d65193b73946ab5d2fc8b281ad26e1cd7c41af6230ac"} Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.317112 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c5404d5ea89531d44b1d65193b73946ab5d2fc8b281ad26e1cd7c41af6230ac" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.335414 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fb66-account-create-q85l7"] Oct 02 09:58:43 crc kubenswrapper[4771]: W1002 09:58:43.352949 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ba7c27c_2a56_4f63_9082_80935e39a0e2.slice/crio-f98eb7f8214f1e253ac1aa886f4b5a04ca9c57a6bbba58ea374cab9e33d05109 WatchSource:0}: Error finding container f98eb7f8214f1e253ac1aa886f4b5a04ca9c57a6bbba58ea374cab9e33d05109: Status 404 returned error can't find the container with id f98eb7f8214f1e253ac1aa886f4b5a04ca9c57a6bbba58ea374cab9e33d05109 Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.380698 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8507c3c-8ba8-47e3-9529-503679e4ffa7-kube-api-access-9wg6m" (OuterVolumeSpecName: "kube-api-access-9wg6m") pod "e8507c3c-8ba8-47e3-9529-503679e4ffa7" (UID: "e8507c3c-8ba8-47e3-9529-503679e4ffa7"). InnerVolumeSpecName "kube-api-access-9wg6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.420321 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wg6m\" (UniqueName: \"kubernetes.io/projected/e8507c3c-8ba8-47e3-9529-503679e4ffa7-kube-api-access-9wg6m\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.499224 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.624519 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-log-ovn\") pod \"514ca08b-64a5-457d-ac4f-e763ab9ad718\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.624960 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/514ca08b-64a5-457d-ac4f-e763ab9ad718-additional-scripts\") pod \"514ca08b-64a5-457d-ac4f-e763ab9ad718\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.625045 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/514ca08b-64a5-457d-ac4f-e763ab9ad718-scripts\") pod \"514ca08b-64a5-457d-ac4f-e763ab9ad718\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.625148 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-run-ovn\") pod \"514ca08b-64a5-457d-ac4f-e763ab9ad718\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.625263 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-run\") pod \"514ca08b-64a5-457d-ac4f-e763ab9ad718\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.625306 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss7pl\" (UniqueName: \"kubernetes.io/projected/514ca08b-64a5-457d-ac4f-e763ab9ad718-kube-api-access-ss7pl\") pod \"514ca08b-64a5-457d-ac4f-e763ab9ad718\" (UID: \"514ca08b-64a5-457d-ac4f-e763ab9ad718\") " Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.624861 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "514ca08b-64a5-457d-ac4f-e763ab9ad718" (UID: "514ca08b-64a5-457d-ac4f-e763ab9ad718"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.626292 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "514ca08b-64a5-457d-ac4f-e763ab9ad718" (UID: "514ca08b-64a5-457d-ac4f-e763ab9ad718"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.626852 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-run" (OuterVolumeSpecName: "var-run") pod "514ca08b-64a5-457d-ac4f-e763ab9ad718" (UID: "514ca08b-64a5-457d-ac4f-e763ab9ad718"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.627213 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/514ca08b-64a5-457d-ac4f-e763ab9ad718-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "514ca08b-64a5-457d-ac4f-e763ab9ad718" (UID: "514ca08b-64a5-457d-ac4f-e763ab9ad718"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.627456 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/514ca08b-64a5-457d-ac4f-e763ab9ad718-scripts" (OuterVolumeSpecName: "scripts") pod "514ca08b-64a5-457d-ac4f-e763ab9ad718" (UID: "514ca08b-64a5-457d-ac4f-e763ab9ad718"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.633215 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/514ca08b-64a5-457d-ac4f-e763ab9ad718-kube-api-access-ss7pl" (OuterVolumeSpecName: "kube-api-access-ss7pl") pod "514ca08b-64a5-457d-ac4f-e763ab9ad718" (UID: "514ca08b-64a5-457d-ac4f-e763ab9ad718"). InnerVolumeSpecName "kube-api-access-ss7pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.781497 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:43 crc kubenswrapper[4771]: E1002 09:58:43.792590 4771 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 09:58:43 crc kubenswrapper[4771]: E1002 09:58:43.792632 4771 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 09:58:43 crc kubenswrapper[4771]: E1002 09:58:43.792688 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift podName:4c4deacb-7b2f-49d1-8c7e-b87913f027d1 nodeName:}" failed. No retries permitted until 2025-10-02 09:58:47.792664921 +0000 UTC m=+1315.440349988 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift") pod "swift-storage-0" (UID: "4c4deacb-7b2f-49d1-8c7e-b87913f027d1") : configmap "swift-ring-files" not found Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.795039 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss7pl\" (UniqueName: \"kubernetes.io/projected/514ca08b-64a5-457d-ac4f-e763ab9ad718-kube-api-access-ss7pl\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.795076 4771 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.795092 4771 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/514ca08b-64a5-457d-ac4f-e763ab9ad718-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.795106 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/514ca08b-64a5-457d-ac4f-e763ab9ad718-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.795122 4771 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.795152 4771 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/514ca08b-64a5-457d-ac4f-e763ab9ad718-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.823478 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-c9bwn"] Oct 02 09:58:43 crc kubenswrapper[4771]: E1002 09:58:43.824024 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8507c3c-8ba8-47e3-9529-503679e4ffa7" containerName="mariadb-account-create" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.824055 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8507c3c-8ba8-47e3-9529-503679e4ffa7" containerName="mariadb-account-create" Oct 02 09:58:43 crc kubenswrapper[4771]: E1002 09:58:43.824119 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="514ca08b-64a5-457d-ac4f-e763ab9ad718" containerName="ovn-config" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.824162 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="514ca08b-64a5-457d-ac4f-e763ab9ad718" containerName="ovn-config" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.824388 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8507c3c-8ba8-47e3-9529-503679e4ffa7" containerName="mariadb-account-create" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.824423 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="514ca08b-64a5-457d-ac4f-e763ab9ad718" containerName="ovn-config" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.825357 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.827986 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.839520 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.840504 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.840652 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.841676 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-c9bwn"] Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.897530 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/087b7d39-d0cf-4896-8220-d79e94810ff5-scripts\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.897611 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f66lq\" (UniqueName: \"kubernetes.io/projected/087b7d39-d0cf-4896-8220-d79e94810ff5-kube-api-access-f66lq\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.897659 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/087b7d39-d0cf-4896-8220-d79e94810ff5-ring-data-devices\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.897684 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-dispersionconf\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.897772 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-swiftconf\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.897897 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-combined-ca-bundle\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:43 crc kubenswrapper[4771]: I1002 09:58:43.900397 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/087b7d39-d0cf-4896-8220-d79e94810ff5-etc-swift\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.002143 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f66lq\" (UniqueName: \"kubernetes.io/projected/087b7d39-d0cf-4896-8220-d79e94810ff5-kube-api-access-f66lq\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.002242 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/087b7d39-d0cf-4896-8220-d79e94810ff5-ring-data-devices\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.002285 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-dispersionconf\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.002344 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-swiftconf\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.002423 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-combined-ca-bundle\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.002505 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/087b7d39-d0cf-4896-8220-d79e94810ff5-etc-swift\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.002560 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/087b7d39-d0cf-4896-8220-d79e94810ff5-scripts\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.003236 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/087b7d39-d0cf-4896-8220-d79e94810ff5-etc-swift\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.003609 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/087b7d39-d0cf-4896-8220-d79e94810ff5-ring-data-devices\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.004483 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/087b7d39-d0cf-4896-8220-d79e94810ff5-scripts\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.006953 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-swiftconf\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.015036 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-dispersionconf\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.015833 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-combined-ca-bundle\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.022340 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f66lq\" (UniqueName: \"kubernetes.io/projected/087b7d39-d0cf-4896-8220-d79e94810ff5-kube-api-access-f66lq\") pod \"swift-ring-rebalance-c9bwn\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.171847 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.313810 4771 generic.go:334] "Generic (PLEG): container finished" podID="1ba7c27c-2a56-4f63-9082-80935e39a0e2" containerID="402479870ae207145b71749e8a1a4d8700d3eeeaf425a34cc6482e569beccf0c" exitCode=0 Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.313894 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fb66-account-create-q85l7" event={"ID":"1ba7c27c-2a56-4f63-9082-80935e39a0e2","Type":"ContainerDied","Data":"402479870ae207145b71749e8a1a4d8700d3eeeaf425a34cc6482e569beccf0c"} Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.313965 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fb66-account-create-q85l7" event={"ID":"1ba7c27c-2a56-4f63-9082-80935e39a0e2","Type":"ContainerStarted","Data":"f98eb7f8214f1e253ac1aa886f4b5a04ca9c57a6bbba58ea374cab9e33d05109"} Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.316876 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5jf58-config-4rvzp" event={"ID":"514ca08b-64a5-457d-ac4f-e763ab9ad718","Type":"ContainerDied","Data":"404f77d21985e3ea26d86aded5b1ee628311bb987be55bed87d685271f5282bb"} Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.316931 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="404f77d21985e3ea26d86aded5b1ee628311bb987be55bed87d685271f5282bb" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.316903 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5jf58-config-4rvzp" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.647467 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-5jf58-config-4rvzp"] Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.657947 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-5jf58-config-4rvzp"] Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.843457 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5jf58-config-8l996"] Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.845153 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.849040 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.863304 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5jf58-config-8l996"] Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.925343 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-log-ovn\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.925440 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e5ddd47f-08a4-4782-b78e-bff1964eaf83-additional-scripts\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.925492 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t98m\" (UniqueName: \"kubernetes.io/projected/e5ddd47f-08a4-4782-b78e-bff1964eaf83-kube-api-access-7t98m\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.925586 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-run\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.925691 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-run-ovn\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:44 crc kubenswrapper[4771]: I1002 09:58:44.925721 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5ddd47f-08a4-4782-b78e-bff1964eaf83-scripts\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.028231 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5ddd47f-08a4-4782-b78e-bff1964eaf83-scripts\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.028448 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-log-ovn\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.028504 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e5ddd47f-08a4-4782-b78e-bff1964eaf83-additional-scripts\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.028559 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t98m\" (UniqueName: \"kubernetes.io/projected/e5ddd47f-08a4-4782-b78e-bff1964eaf83-kube-api-access-7t98m\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.028655 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-run\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.028770 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-run-ovn\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.028986 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-run-ovn\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.028981 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-log-ovn\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.029509 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-run\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.029776 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e5ddd47f-08a4-4782-b78e-bff1964eaf83-additional-scripts\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.030388 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5ddd47f-08a4-4782-b78e-bff1964eaf83-scripts\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.067201 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t98m\" (UniqueName: \"kubernetes.io/projected/e5ddd47f-08a4-4782-b78e-bff1964eaf83-kube-api-access-7t98m\") pod \"ovn-controller-5jf58-config-8l996\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.163474 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:45 crc kubenswrapper[4771]: I1002 09:58:45.704134 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="514ca08b-64a5-457d-ac4f-e763ab9ad718" path="/var/lib/kubelet/pods/514ca08b-64a5-457d-ac4f-e763ab9ad718/volumes" Oct 02 09:58:46 crc kubenswrapper[4771]: I1002 09:58:46.370652 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f944-account-create-752ln"] Oct 02 09:58:46 crc kubenswrapper[4771]: I1002 09:58:46.373548 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f944-account-create-752ln" Oct 02 09:58:46 crc kubenswrapper[4771]: I1002 09:58:46.478701 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 02 09:58:46 crc kubenswrapper[4771]: I1002 09:58:46.500140 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f944-account-create-752ln"] Oct 02 09:58:46 crc kubenswrapper[4771]: I1002 09:58:46.584435 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhczk\" (UniqueName: \"kubernetes.io/projected/02570cd0-0131-410c-9053-8909d7b5d95d-kube-api-access-xhczk\") pod \"keystone-f944-account-create-752ln\" (UID: \"02570cd0-0131-410c-9053-8909d7b5d95d\") " pod="openstack/keystone-f944-account-create-752ln" Oct 02 09:58:46 crc kubenswrapper[4771]: I1002 09:58:46.685898 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhczk\" (UniqueName: \"kubernetes.io/projected/02570cd0-0131-410c-9053-8909d7b5d95d-kube-api-access-xhczk\") pod \"keystone-f944-account-create-752ln\" (UID: \"02570cd0-0131-410c-9053-8909d7b5d95d\") " pod="openstack/keystone-f944-account-create-752ln" Oct 02 09:58:46 crc kubenswrapper[4771]: I1002 09:58:46.718219 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhczk\" (UniqueName: \"kubernetes.io/projected/02570cd0-0131-410c-9053-8909d7b5d95d-kube-api-access-xhczk\") pod \"keystone-f944-account-create-752ln\" (UID: \"02570cd0-0131-410c-9053-8909d7b5d95d\") " pod="openstack/keystone-f944-account-create-752ln" Oct 02 09:58:46 crc kubenswrapper[4771]: I1002 09:58:46.742570 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-11a9-account-create-4nhgk"] Oct 02 09:58:46 crc kubenswrapper[4771]: I1002 09:58:46.745832 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-11a9-account-create-4nhgk" Oct 02 09:58:46 crc kubenswrapper[4771]: I1002 09:58:46.750329 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 02 09:58:46 crc kubenswrapper[4771]: I1002 09:58:46.785120 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-11a9-account-create-4nhgk"] Oct 02 09:58:46 crc kubenswrapper[4771]: I1002 09:58:46.799400 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f944-account-create-752ln" Oct 02 09:58:46 crc kubenswrapper[4771]: I1002 09:58:46.898086 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hghjg\" (UniqueName: \"kubernetes.io/projected/97d628d6-bb40-4813-bea2-4bd87d549699-kube-api-access-hghjg\") pod \"placement-11a9-account-create-4nhgk\" (UID: \"97d628d6-bb40-4813-bea2-4bd87d549699\") " pod="openstack/placement-11a9-account-create-4nhgk" Oct 02 09:58:47 crc kubenswrapper[4771]: I1002 09:58:47.001068 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hghjg\" (UniqueName: \"kubernetes.io/projected/97d628d6-bb40-4813-bea2-4bd87d549699-kube-api-access-hghjg\") pod \"placement-11a9-account-create-4nhgk\" (UID: \"97d628d6-bb40-4813-bea2-4bd87d549699\") " pod="openstack/placement-11a9-account-create-4nhgk" Oct 02 09:58:47 crc kubenswrapper[4771]: I1002 09:58:47.026165 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hghjg\" (UniqueName: \"kubernetes.io/projected/97d628d6-bb40-4813-bea2-4bd87d549699-kube-api-access-hghjg\") pod \"placement-11a9-account-create-4nhgk\" (UID: \"97d628d6-bb40-4813-bea2-4bd87d549699\") " pod="openstack/placement-11a9-account-create-4nhgk" Oct 02 09:58:47 crc kubenswrapper[4771]: I1002 09:58:47.107535 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-11a9-account-create-4nhgk" Oct 02 09:58:47 crc kubenswrapper[4771]: I1002 09:58:47.238259 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fb66-account-create-q85l7" Oct 02 09:58:47 crc kubenswrapper[4771]: I1002 09:58:47.365722 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fb66-account-create-q85l7" event={"ID":"1ba7c27c-2a56-4f63-9082-80935e39a0e2","Type":"ContainerDied","Data":"f98eb7f8214f1e253ac1aa886f4b5a04ca9c57a6bbba58ea374cab9e33d05109"} Oct 02 09:58:47 crc kubenswrapper[4771]: I1002 09:58:47.365768 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f98eb7f8214f1e253ac1aa886f4b5a04ca9c57a6bbba58ea374cab9e33d05109" Oct 02 09:58:47 crc kubenswrapper[4771]: I1002 09:58:47.365829 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fb66-account-create-q85l7" Oct 02 09:58:47 crc kubenswrapper[4771]: I1002 09:58:47.412707 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjk5k\" (UniqueName: \"kubernetes.io/projected/1ba7c27c-2a56-4f63-9082-80935e39a0e2-kube-api-access-pjk5k\") pod \"1ba7c27c-2a56-4f63-9082-80935e39a0e2\" (UID: \"1ba7c27c-2a56-4f63-9082-80935e39a0e2\") " Oct 02 09:58:47 crc kubenswrapper[4771]: I1002 09:58:47.429678 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ba7c27c-2a56-4f63-9082-80935e39a0e2-kube-api-access-pjk5k" (OuterVolumeSpecName: "kube-api-access-pjk5k") pod "1ba7c27c-2a56-4f63-9082-80935e39a0e2" (UID: "1ba7c27c-2a56-4f63-9082-80935e39a0e2"). InnerVolumeSpecName "kube-api-access-pjk5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:47 crc kubenswrapper[4771]: I1002 09:58:47.517792 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjk5k\" (UniqueName: \"kubernetes.io/projected/1ba7c27c-2a56-4f63-9082-80935e39a0e2-kube-api-access-pjk5k\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:47 crc kubenswrapper[4771]: I1002 09:58:47.835883 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:47 crc kubenswrapper[4771]: E1002 09:58:47.838037 4771 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 09:58:47 crc kubenswrapper[4771]: E1002 09:58:47.838072 4771 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 09:58:47 crc kubenswrapper[4771]: E1002 09:58:47.838148 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift podName:4c4deacb-7b2f-49d1-8c7e-b87913f027d1 nodeName:}" failed. No retries permitted until 2025-10-02 09:58:55.838106776 +0000 UTC m=+1323.485792053 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift") pod "swift-storage-0" (UID: "4c4deacb-7b2f-49d1-8c7e-b87913f027d1") : configmap "swift-ring-files" not found Oct 02 09:58:47 crc kubenswrapper[4771]: I1002 09:58:47.927728 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-11a9-account-create-4nhgk"] Oct 02 09:58:47 crc kubenswrapper[4771]: W1002 09:58:47.937742 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97d628d6_bb40_4813_bea2_4bd87d549699.slice/crio-44caf7f16565ba2ca246174980e0166180d4b8392b9109c3588a678148448863 WatchSource:0}: Error finding container 44caf7f16565ba2ca246174980e0166180d4b8392b9109c3588a678148448863: Status 404 returned error can't find the container with id 44caf7f16565ba2ca246174980e0166180d4b8392b9109c3588a678148448863 Oct 02 09:58:48 crc kubenswrapper[4771]: I1002 09:58:48.314229 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f944-account-create-752ln"] Oct 02 09:58:48 crc kubenswrapper[4771]: W1002 09:58:48.324743 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02570cd0_0131_410c_9053_8909d7b5d95d.slice/crio-7bc8f61ec62778c974086d54f8dff8909b2d10f3eff6a5bb2e799ea3ca86c848 WatchSource:0}: Error finding container 7bc8f61ec62778c974086d54f8dff8909b2d10f3eff6a5bb2e799ea3ca86c848: Status 404 returned error can't find the container with id 7bc8f61ec62778c974086d54f8dff8909b2d10f3eff6a5bb2e799ea3ca86c848 Oct 02 09:58:48 crc kubenswrapper[4771]: I1002 09:58:48.341792 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-c9bwn"] Oct 02 09:58:48 crc kubenswrapper[4771]: I1002 09:58:48.351267 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:58:48 crc kubenswrapper[4771]: I1002 09:58:48.381924 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5jf58-config-8l996"] Oct 02 09:58:48 crc kubenswrapper[4771]: I1002 09:58:48.402900 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"95cd3d5a-1000-4212-b4d9-5d0da549c7f3","Type":"ContainerStarted","Data":"c3286830dd0529949a1d53aeff88ff2c8e1154091dff38d555badaf1e6af5d75"} Oct 02 09:58:48 crc kubenswrapper[4771]: I1002 09:58:48.413444 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f944-account-create-752ln" event={"ID":"02570cd0-0131-410c-9053-8909d7b5d95d","Type":"ContainerStarted","Data":"7bc8f61ec62778c974086d54f8dff8909b2d10f3eff6a5bb2e799ea3ca86c848"} Oct 02 09:58:48 crc kubenswrapper[4771]: I1002 09:58:48.416466 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-c9bwn" event={"ID":"087b7d39-d0cf-4896-8220-d79e94810ff5","Type":"ContainerStarted","Data":"5722c920dc00e8b6342fa579faccc501881e9f5a0edb9545ce3c4e5eed845850"} Oct 02 09:58:48 crc kubenswrapper[4771]: I1002 09:58:48.418826 4771 generic.go:334] "Generic (PLEG): container finished" podID="97d628d6-bb40-4813-bea2-4bd87d549699" containerID="544b74029917cc328962e02ba28141d9fca746ebba4a1d968bd3847fcef3e317" exitCode=0 Oct 02 09:58:48 crc kubenswrapper[4771]: I1002 09:58:48.418864 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-11a9-account-create-4nhgk" event={"ID":"97d628d6-bb40-4813-bea2-4bd87d549699","Type":"ContainerDied","Data":"544b74029917cc328962e02ba28141d9fca746ebba4a1d968bd3847fcef3e317"} Oct 02 09:58:48 crc kubenswrapper[4771]: I1002 09:58:48.418890 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-11a9-account-create-4nhgk" event={"ID":"97d628d6-bb40-4813-bea2-4bd87d549699","Type":"ContainerStarted","Data":"44caf7f16565ba2ca246174980e0166180d4b8392b9109c3588a678148448863"} Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.041739 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-sd45z"] Oct 02 09:58:49 crc kubenswrapper[4771]: E1002 09:58:49.042727 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba7c27c-2a56-4f63-9082-80935e39a0e2" containerName="mariadb-account-create" Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.042761 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba7c27c-2a56-4f63-9082-80935e39a0e2" containerName="mariadb-account-create" Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.043028 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ba7c27c-2a56-4f63-9082-80935e39a0e2" containerName="mariadb-account-create" Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.043943 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sd45z" Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.066540 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp6qj\" (UniqueName: \"kubernetes.io/projected/c8c508c8-a5cb-4b9a-866f-c695fd9f1d02-kube-api-access-vp6qj\") pod \"mysqld-exporter-openstack-cell1-db-create-sd45z\" (UID: \"c8c508c8-a5cb-4b9a-866f-c695fd9f1d02\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-sd45z" Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.104601 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.130279 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-sd45z"] Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.169745 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp6qj\" (UniqueName: \"kubernetes.io/projected/c8c508c8-a5cb-4b9a-866f-c695fd9f1d02-kube-api-access-vp6qj\") pod \"mysqld-exporter-openstack-cell1-db-create-sd45z\" (UID: \"c8c508c8-a5cb-4b9a-866f-c695fd9f1d02\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-sd45z" Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.218774 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp6qj\" (UniqueName: \"kubernetes.io/projected/c8c508c8-a5cb-4b9a-866f-c695fd9f1d02-kube-api-access-vp6qj\") pod \"mysqld-exporter-openstack-cell1-db-create-sd45z\" (UID: \"c8c508c8-a5cb-4b9a-866f-c695fd9f1d02\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-sd45z" Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.242446 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-cwm9n"] Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.242697 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" podUID="7eb14f25-cd51-4f54-ac31-3eb9cf51e152" containerName="dnsmasq-dns" containerID="cri-o://efa3ef697839c4982a1936f5891ff9ca59a9e10a47012a890cba3c5167d8ec50" gracePeriod=10 Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.370101 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sd45z" Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.469034 4771 generic.go:334] "Generic (PLEG): container finished" podID="7eb14f25-cd51-4f54-ac31-3eb9cf51e152" containerID="efa3ef697839c4982a1936f5891ff9ca59a9e10a47012a890cba3c5167d8ec50" exitCode=0 Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.469252 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" event={"ID":"7eb14f25-cd51-4f54-ac31-3eb9cf51e152","Type":"ContainerDied","Data":"efa3ef697839c4982a1936f5891ff9ca59a9e10a47012a890cba3c5167d8ec50"} Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.488013 4771 generic.go:334] "Generic (PLEG): container finished" podID="02570cd0-0131-410c-9053-8909d7b5d95d" containerID="d696f2e8df0ce0072ce0fb694fb19ee381869cdff489626a5ee7d213485a4e4c" exitCode=0 Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.488169 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f944-account-create-752ln" event={"ID":"02570cd0-0131-410c-9053-8909d7b5d95d","Type":"ContainerDied","Data":"d696f2e8df0ce0072ce0fb694fb19ee381869cdff489626a5ee7d213485a4e4c"} Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.494880 4771 generic.go:334] "Generic (PLEG): container finished" podID="e5ddd47f-08a4-4782-b78e-bff1964eaf83" containerID="4ff2ad23fc2d0fe2e7d27804d86888db3fdce232cd0c67272337d99907ba4e7c" exitCode=0 Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.495185 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5jf58-config-8l996" event={"ID":"e5ddd47f-08a4-4782-b78e-bff1964eaf83","Type":"ContainerDied","Data":"4ff2ad23fc2d0fe2e7d27804d86888db3fdce232cd0c67272337d99907ba4e7c"} Oct 02 09:58:49 crc kubenswrapper[4771]: I1002 09:58:49.495260 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5jf58-config-8l996" event={"ID":"e5ddd47f-08a4-4782-b78e-bff1964eaf83","Type":"ContainerStarted","Data":"e61a09c608283553dada646923c85b38166b3dcfb04b9cc54306f5b13dab22ff"} Oct 02 09:58:50 crc kubenswrapper[4771]: I1002 09:58:50.450511 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-sd45z"] Oct 02 09:58:50 crc kubenswrapper[4771]: I1002 09:58:50.515784 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" event={"ID":"7eb14f25-cd51-4f54-ac31-3eb9cf51e152","Type":"ContainerDied","Data":"0ed8aafc01642f615bb3a2761489092d237108ea4159b07d34e646d67226ad5f"} Oct 02 09:58:50 crc kubenswrapper[4771]: I1002 09:58:50.515838 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ed8aafc01642f615bb3a2761489092d237108ea4159b07d34e646d67226ad5f" Oct 02 09:58:50 crc kubenswrapper[4771]: I1002 09:58:50.522688 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sd45z" event={"ID":"c8c508c8-a5cb-4b9a-866f-c695fd9f1d02","Type":"ContainerStarted","Data":"21bc6c6cfe37164e19cbdd1ef59c136b3c1eb759c80d728d131624b68c18f7af"} Oct 02 09:58:50 crc kubenswrapper[4771]: I1002 09:58:50.550499 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:50 crc kubenswrapper[4771]: I1002 09:58:50.714534 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-ovsdbserver-sb\") pod \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " Oct 02 09:58:50 crc kubenswrapper[4771]: I1002 09:58:50.714613 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-config\") pod \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " Oct 02 09:58:50 crc kubenswrapper[4771]: I1002 09:58:50.714663 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-dns-svc\") pod \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " Oct 02 09:58:50 crc kubenswrapper[4771]: I1002 09:58:50.714845 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp9ww\" (UniqueName: \"kubernetes.io/projected/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-kube-api-access-hp9ww\") pod \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " Oct 02 09:58:50 crc kubenswrapper[4771]: I1002 09:58:50.714964 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-ovsdbserver-nb\") pod \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\" (UID: \"7eb14f25-cd51-4f54-ac31-3eb9cf51e152\") " Oct 02 09:58:50 crc kubenswrapper[4771]: I1002 09:58:50.776043 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-kube-api-access-hp9ww" (OuterVolumeSpecName: "kube-api-access-hp9ww") pod "7eb14f25-cd51-4f54-ac31-3eb9cf51e152" (UID: "7eb14f25-cd51-4f54-ac31-3eb9cf51e152"). InnerVolumeSpecName "kube-api-access-hp9ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:50 crc kubenswrapper[4771]: I1002 09:58:50.818730 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp9ww\" (UniqueName: \"kubernetes.io/projected/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-kube-api-access-hp9ww\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:50 crc kubenswrapper[4771]: I1002 09:58:50.958666 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7eb14f25-cd51-4f54-ac31-3eb9cf51e152" (UID: "7eb14f25-cd51-4f54-ac31-3eb9cf51e152"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.032375 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.183379 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7eb14f25-cd51-4f54-ac31-3eb9cf51e152" (UID: "7eb14f25-cd51-4f54-ac31-3eb9cf51e152"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.244625 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.338305 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-config" (OuterVolumeSpecName: "config") pod "7eb14f25-cd51-4f54-ac31-3eb9cf51e152" (UID: "7eb14f25-cd51-4f54-ac31-3eb9cf51e152"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.354349 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.394016 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7eb14f25-cd51-4f54-ac31-3eb9cf51e152" (UID: "7eb14f25-cd51-4f54-ac31-3eb9cf51e152"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.455872 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7eb14f25-cd51-4f54-ac31-3eb9cf51e152-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.536977 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f944-account-create-752ln" event={"ID":"02570cd0-0131-410c-9053-8909d7b5d95d","Type":"ContainerDied","Data":"7bc8f61ec62778c974086d54f8dff8909b2d10f3eff6a5bb2e799ea3ca86c848"} Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.537023 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bc8f61ec62778c974086d54f8dff8909b2d10f3eff6a5bb2e799ea3ca86c848" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.539401 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5jf58-config-8l996" event={"ID":"e5ddd47f-08a4-4782-b78e-bff1964eaf83","Type":"ContainerDied","Data":"e61a09c608283553dada646923c85b38166b3dcfb04b9cc54306f5b13dab22ff"} Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.539429 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e61a09c608283553dada646923c85b38166b3dcfb04b9cc54306f5b13dab22ff" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.542109 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-11a9-account-create-4nhgk" event={"ID":"97d628d6-bb40-4813-bea2-4bd87d549699","Type":"ContainerDied","Data":"44caf7f16565ba2ca246174980e0166180d4b8392b9109c3588a678148448863"} Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.542189 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44caf7f16565ba2ca246174980e0166180d4b8392b9109c3588a678148448863" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.542153 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.809573 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-11a9-account-create-4nhgk" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.842635 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f944-account-create-752ln" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.864439 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hghjg\" (UniqueName: \"kubernetes.io/projected/97d628d6-bb40-4813-bea2-4bd87d549699-kube-api-access-hghjg\") pod \"97d628d6-bb40-4813-bea2-4bd87d549699\" (UID: \"97d628d6-bb40-4813-bea2-4bd87d549699\") " Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.864585 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhczk\" (UniqueName: \"kubernetes.io/projected/02570cd0-0131-410c-9053-8909d7b5d95d-kube-api-access-xhczk\") pod \"02570cd0-0131-410c-9053-8909d7b5d95d\" (UID: \"02570cd0-0131-410c-9053-8909d7b5d95d\") " Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.879711 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97d628d6-bb40-4813-bea2-4bd87d549699-kube-api-access-hghjg" (OuterVolumeSpecName: "kube-api-access-hghjg") pod "97d628d6-bb40-4813-bea2-4bd87d549699" (UID: "97d628d6-bb40-4813-bea2-4bd87d549699"). InnerVolumeSpecName "kube-api-access-hghjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.880012 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.882196 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02570cd0-0131-410c-9053-8909d7b5d95d-kube-api-access-xhczk" (OuterVolumeSpecName: "kube-api-access-xhczk") pod "02570cd0-0131-410c-9053-8909d7b5d95d" (UID: "02570cd0-0131-410c-9053-8909d7b5d95d"). InnerVolumeSpecName "kube-api-access-xhczk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.967241 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-run\") pod \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.967332 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t98m\" (UniqueName: \"kubernetes.io/projected/e5ddd47f-08a4-4782-b78e-bff1964eaf83-kube-api-access-7t98m\") pod \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.967408 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e5ddd47f-08a4-4782-b78e-bff1964eaf83-additional-scripts\") pod \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.967460 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-run-ovn\") pod \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.967453 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-run" (OuterVolumeSpecName: "var-run") pod "e5ddd47f-08a4-4782-b78e-bff1964eaf83" (UID: "e5ddd47f-08a4-4782-b78e-bff1964eaf83"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.967560 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5ddd47f-08a4-4782-b78e-bff1964eaf83-scripts\") pod \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.967617 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e5ddd47f-08a4-4782-b78e-bff1964eaf83" (UID: "e5ddd47f-08a4-4782-b78e-bff1964eaf83"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.967628 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-log-ovn\") pod \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\" (UID: \"e5ddd47f-08a4-4782-b78e-bff1964eaf83\") " Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.967682 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e5ddd47f-08a4-4782-b78e-bff1964eaf83" (UID: "e5ddd47f-08a4-4782-b78e-bff1964eaf83"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.968683 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5ddd47f-08a4-4782-b78e-bff1964eaf83-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e5ddd47f-08a4-4782-b78e-bff1964eaf83" (UID: "e5ddd47f-08a4-4782-b78e-bff1964eaf83"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.969006 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5ddd47f-08a4-4782-b78e-bff1964eaf83-scripts" (OuterVolumeSpecName: "scripts") pod "e5ddd47f-08a4-4782-b78e-bff1964eaf83" (UID: "e5ddd47f-08a4-4782-b78e-bff1964eaf83"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.970007 4771 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.970031 4771 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e5ddd47f-08a4-4782-b78e-bff1964eaf83-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.970050 4771 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.970064 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5ddd47f-08a4-4782-b78e-bff1964eaf83-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.970076 4771 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e5ddd47f-08a4-4782-b78e-bff1964eaf83-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.970091 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hghjg\" (UniqueName: \"kubernetes.io/projected/97d628d6-bb40-4813-bea2-4bd87d549699-kube-api-access-hghjg\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.970106 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhczk\" (UniqueName: \"kubernetes.io/projected/02570cd0-0131-410c-9053-8909d7b5d95d-kube-api-access-xhczk\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:51 crc kubenswrapper[4771]: I1002 09:58:51.973635 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5ddd47f-08a4-4782-b78e-bff1964eaf83-kube-api-access-7t98m" (OuterVolumeSpecName: "kube-api-access-7t98m") pod "e5ddd47f-08a4-4782-b78e-bff1964eaf83" (UID: "e5ddd47f-08a4-4782-b78e-bff1964eaf83"). InnerVolumeSpecName "kube-api-access-7t98m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.072118 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t98m\" (UniqueName: \"kubernetes.io/projected/e5ddd47f-08a4-4782-b78e-bff1964eaf83-kube-api-access-7t98m\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.208608 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-t42dx"] Oct 02 09:58:52 crc kubenswrapper[4771]: E1002 09:58:52.209212 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eb14f25-cd51-4f54-ac31-3eb9cf51e152" containerName="init" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.209237 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eb14f25-cd51-4f54-ac31-3eb9cf51e152" containerName="init" Oct 02 09:58:52 crc kubenswrapper[4771]: E1002 09:58:52.209258 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02570cd0-0131-410c-9053-8909d7b5d95d" containerName="mariadb-account-create" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.209267 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="02570cd0-0131-410c-9053-8909d7b5d95d" containerName="mariadb-account-create" Oct 02 09:58:52 crc kubenswrapper[4771]: E1002 09:58:52.209284 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5ddd47f-08a4-4782-b78e-bff1964eaf83" containerName="ovn-config" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.209292 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5ddd47f-08a4-4782-b78e-bff1964eaf83" containerName="ovn-config" Oct 02 09:58:52 crc kubenswrapper[4771]: E1002 09:58:52.209306 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97d628d6-bb40-4813-bea2-4bd87d549699" containerName="mariadb-account-create" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.209313 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="97d628d6-bb40-4813-bea2-4bd87d549699" containerName="mariadb-account-create" Oct 02 09:58:52 crc kubenswrapper[4771]: E1002 09:58:52.209346 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eb14f25-cd51-4f54-ac31-3eb9cf51e152" containerName="dnsmasq-dns" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.209355 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eb14f25-cd51-4f54-ac31-3eb9cf51e152" containerName="dnsmasq-dns" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.209622 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="02570cd0-0131-410c-9053-8909d7b5d95d" containerName="mariadb-account-create" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.209646 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eb14f25-cd51-4f54-ac31-3eb9cf51e152" containerName="dnsmasq-dns" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.209671 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="97d628d6-bb40-4813-bea2-4bd87d549699" containerName="mariadb-account-create" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.209690 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5ddd47f-08a4-4782-b78e-bff1964eaf83" containerName="ovn-config" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.218951 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.223296 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lnfgz" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.223636 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.239045 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-t42dx"] Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.277321 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-config-data\") pod \"glance-db-sync-t42dx\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.277557 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx4jk\" (UniqueName: \"kubernetes.io/projected/4c6203ef-b851-49ba-ae59-9b62c473c497-kube-api-access-rx4jk\") pod \"glance-db-sync-t42dx\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.277939 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-db-sync-config-data\") pod \"glance-db-sync-t42dx\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.278066 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-combined-ca-bundle\") pod \"glance-db-sync-t42dx\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.380061 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx4jk\" (UniqueName: \"kubernetes.io/projected/4c6203ef-b851-49ba-ae59-9b62c473c497-kube-api-access-rx4jk\") pod \"glance-db-sync-t42dx\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.380466 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-db-sync-config-data\") pod \"glance-db-sync-t42dx\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.380551 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-combined-ca-bundle\") pod \"glance-db-sync-t42dx\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.380765 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-config-data\") pod \"glance-db-sync-t42dx\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.385198 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-db-sync-config-data\") pod \"glance-db-sync-t42dx\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.385725 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-config-data\") pod \"glance-db-sync-t42dx\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.389719 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-combined-ca-bundle\") pod \"glance-db-sync-t42dx\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.420240 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx4jk\" (UniqueName: \"kubernetes.io/projected/4c6203ef-b851-49ba-ae59-9b62c473c497-kube-api-access-rx4jk\") pod \"glance-db-sync-t42dx\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.556582 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"95cd3d5a-1000-4212-b4d9-5d0da549c7f3","Type":"ContainerStarted","Data":"7e88704fb0629cfd87c9422d98c45c8ecb8684b1aa2ddf1d73efee5edce3e523"} Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.559523 4771 generic.go:334] "Generic (PLEG): container finished" podID="c8c508c8-a5cb-4b9a-866f-c695fd9f1d02" containerID="4c947da2879f48e37fb0dbc90a4af1dea5148f66c286d4eb6382e0387f8ae9de" exitCode=0 Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.559595 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5jf58-config-8l996" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.559662 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sd45z" event={"ID":"c8c508c8-a5cb-4b9a-866f-c695fd9f1d02","Type":"ContainerDied","Data":"4c947da2879f48e37fb0dbc90a4af1dea5148f66c286d4eb6382e0387f8ae9de"} Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.559749 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f944-account-create-752ln" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.559814 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-11a9-account-create-4nhgk" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.584874 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-t42dx" Oct 02 09:58:52 crc kubenswrapper[4771]: I1002 09:58:52.992895 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-5jf58-config-8l996"] Oct 02 09:58:53 crc kubenswrapper[4771]: I1002 09:58:53.006323 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-5jf58-config-8l996"] Oct 02 09:58:53 crc kubenswrapper[4771]: I1002 09:58:53.428248 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="b7dad196-4ebc-43e0-b850-0820f2b072c9" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.131:5671: connect: connection refused" Oct 02 09:58:53 crc kubenswrapper[4771]: I1002 09:58:53.709424 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5ddd47f-08a4-4782-b78e-bff1964eaf83" path="/var/lib/kubelet/pods/e5ddd47f-08a4-4782-b78e-bff1964eaf83/volumes" Oct 02 09:58:53 crc kubenswrapper[4771]: I1002 09:58:53.923580 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="ebc075f5-e2ba-4789-a38e-4065fe661ada" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.132:5671: connect: connection refused" Oct 02 09:58:55 crc kubenswrapper[4771]: I1002 09:58:55.291565 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sd45z" Oct 02 09:58:55 crc kubenswrapper[4771]: I1002 09:58:55.353470 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp6qj\" (UniqueName: \"kubernetes.io/projected/c8c508c8-a5cb-4b9a-866f-c695fd9f1d02-kube-api-access-vp6qj\") pod \"c8c508c8-a5cb-4b9a-866f-c695fd9f1d02\" (UID: \"c8c508c8-a5cb-4b9a-866f-c695fd9f1d02\") " Oct 02 09:58:55 crc kubenswrapper[4771]: I1002 09:58:55.384147 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8c508c8-a5cb-4b9a-866f-c695fd9f1d02-kube-api-access-vp6qj" (OuterVolumeSpecName: "kube-api-access-vp6qj") pod "c8c508c8-a5cb-4b9a-866f-c695fd9f1d02" (UID: "c8c508c8-a5cb-4b9a-866f-c695fd9f1d02"). InnerVolumeSpecName "kube-api-access-vp6qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:58:55 crc kubenswrapper[4771]: I1002 09:58:55.461458 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp6qj\" (UniqueName: \"kubernetes.io/projected/c8c508c8-a5cb-4b9a-866f-c695fd9f1d02-kube-api-access-vp6qj\") on node \"crc\" DevicePath \"\"" Oct 02 09:58:55 crc kubenswrapper[4771]: I1002 09:58:55.592653 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sd45z" event={"ID":"c8c508c8-a5cb-4b9a-866f-c695fd9f1d02","Type":"ContainerDied","Data":"21bc6c6cfe37164e19cbdd1ef59c136b3c1eb759c80d728d131624b68c18f7af"} Oct 02 09:58:55 crc kubenswrapper[4771]: I1002 09:58:55.592717 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21bc6c6cfe37164e19cbdd1ef59c136b3c1eb759c80d728d131624b68c18f7af" Oct 02 09:58:55 crc kubenswrapper[4771]: I1002 09:58:55.592724 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sd45z" Oct 02 09:58:55 crc kubenswrapper[4771]: I1002 09:58:55.776557 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-t42dx"] Oct 02 09:58:55 crc kubenswrapper[4771]: I1002 09:58:55.871142 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:58:55 crc kubenswrapper[4771]: E1002 09:58:55.871330 4771 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 09:58:55 crc kubenswrapper[4771]: E1002 09:58:55.871569 4771 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 09:58:55 crc kubenswrapper[4771]: E1002 09:58:55.871620 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift podName:4c4deacb-7b2f-49d1-8c7e-b87913f027d1 nodeName:}" failed. No retries permitted until 2025-10-02 09:59:11.871605908 +0000 UTC m=+1339.519290965 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift") pod "swift-storage-0" (UID: "4c4deacb-7b2f-49d1-8c7e-b87913f027d1") : configmap "swift-ring-files" not found Oct 02 09:58:59 crc kubenswrapper[4771]: I1002 09:58:59.370888 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-f206-account-create-nlvdg"] Oct 02 09:58:59 crc kubenswrapper[4771]: E1002 09:58:59.371949 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c508c8-a5cb-4b9a-866f-c695fd9f1d02" containerName="mariadb-database-create" Oct 02 09:58:59 crc kubenswrapper[4771]: I1002 09:58:59.371967 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c508c8-a5cb-4b9a-866f-c695fd9f1d02" containerName="mariadb-database-create" Oct 02 09:58:59 crc kubenswrapper[4771]: I1002 09:58:59.372319 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8c508c8-a5cb-4b9a-866f-c695fd9f1d02" containerName="mariadb-database-create" Oct 02 09:58:59 crc kubenswrapper[4771]: I1002 09:58:59.373307 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-f206-account-create-nlvdg" Oct 02 09:58:59 crc kubenswrapper[4771]: I1002 09:58:59.376332 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Oct 02 09:58:59 crc kubenswrapper[4771]: I1002 09:58:59.382349 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-f206-account-create-nlvdg"] Oct 02 09:58:59 crc kubenswrapper[4771]: I1002 09:58:59.469640 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk5lt\" (UniqueName: \"kubernetes.io/projected/55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc-kube-api-access-xk5lt\") pod \"mysqld-exporter-f206-account-create-nlvdg\" (UID: \"55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc\") " pod="openstack/mysqld-exporter-f206-account-create-nlvdg" Oct 02 09:58:59 crc kubenswrapper[4771]: I1002 09:58:59.571258 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk5lt\" (UniqueName: \"kubernetes.io/projected/55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc-kube-api-access-xk5lt\") pod \"mysqld-exporter-f206-account-create-nlvdg\" (UID: \"55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc\") " pod="openstack/mysqld-exporter-f206-account-create-nlvdg" Oct 02 09:58:59 crc kubenswrapper[4771]: I1002 09:58:59.591844 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk5lt\" (UniqueName: \"kubernetes.io/projected/55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc-kube-api-access-xk5lt\") pod \"mysqld-exporter-f206-account-create-nlvdg\" (UID: \"55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc\") " pod="openstack/mysqld-exporter-f206-account-create-nlvdg" Oct 02 09:58:59 crc kubenswrapper[4771]: I1002 09:58:59.640939 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-t42dx" event={"ID":"4c6203ef-b851-49ba-ae59-9b62c473c497","Type":"ContainerStarted","Data":"1a3e64e0b57fc0ff5ac8675f92580dd7ac4dc3190e3e759bb322291a080db9fc"} Oct 02 09:58:59 crc kubenswrapper[4771]: I1002 09:58:59.740038 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-f206-account-create-nlvdg" Oct 02 09:59:00 crc kubenswrapper[4771]: I1002 09:59:00.259053 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-f206-account-create-nlvdg"] Oct 02 09:59:00 crc kubenswrapper[4771]: I1002 09:59:00.661264 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"95cd3d5a-1000-4212-b4d9-5d0da549c7f3","Type":"ContainerStarted","Data":"61c3b16f18cbb70283e079c3cf0a250091cf91ceb3e801c047cf6918fb937be1"} Oct 02 09:59:00 crc kubenswrapper[4771]: I1002 09:59:00.666748 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-c9bwn" event={"ID":"087b7d39-d0cf-4896-8220-d79e94810ff5","Type":"ContainerStarted","Data":"44a564538fad32d29e4ba361673681ee694bfa50ed1594361eaa7d7e84a46bf3"} Oct 02 09:59:00 crc kubenswrapper[4771]: I1002 09:59:00.669168 4771 generic.go:334] "Generic (PLEG): container finished" podID="55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc" containerID="6cbac7afbdc710086e1ea377ba7123136a800ece8dad7712c4835e4829f8d2bf" exitCode=0 Oct 02 09:59:00 crc kubenswrapper[4771]: I1002 09:59:00.669228 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-f206-account-create-nlvdg" event={"ID":"55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc","Type":"ContainerDied","Data":"6cbac7afbdc710086e1ea377ba7123136a800ece8dad7712c4835e4829f8d2bf"} Oct 02 09:59:00 crc kubenswrapper[4771]: I1002 09:59:00.669265 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-f206-account-create-nlvdg" event={"ID":"55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc","Type":"ContainerStarted","Data":"e6ccd51ddad9f03a61f4694050b8661a57ee5f99376b641807fd92c7f37d3729"} Oct 02 09:59:00 crc kubenswrapper[4771]: I1002 09:59:00.689288 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=4.699476891 podStartE2EDuration="1m32.689265159s" podCreationTimestamp="2025-10-02 09:57:28 +0000 UTC" firstStartedPulling="2025-10-02 09:57:31.562499408 +0000 UTC m=+1239.210184475" lastFinishedPulling="2025-10-02 09:58:59.552287676 +0000 UTC m=+1327.199972743" observedRunningTime="2025-10-02 09:59:00.689004513 +0000 UTC m=+1328.336689580" watchObservedRunningTime="2025-10-02 09:59:00.689265159 +0000 UTC m=+1328.336950226" Oct 02 09:59:00 crc kubenswrapper[4771]: E1002 09:59:00.733998 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55933e6a_efc8_4c5d_9f6a_bf9c7e73ffdc.slice/crio-6cbac7afbdc710086e1ea377ba7123136a800ece8dad7712c4835e4829f8d2bf.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55933e6a_efc8_4c5d_9f6a_bf9c7e73ffdc.slice/crio-conmon-6cbac7afbdc710086e1ea377ba7123136a800ece8dad7712c4835e4829f8d2bf.scope\": RecentStats: unable to find data in memory cache]" Oct 02 09:59:00 crc kubenswrapper[4771]: I1002 09:59:00.747930 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-c9bwn" podStartSLOduration=6.8457981740000005 podStartE2EDuration="17.747902449s" podCreationTimestamp="2025-10-02 09:58:43 +0000 UTC" firstStartedPulling="2025-10-02 09:58:48.350918778 +0000 UTC m=+1315.998603845" lastFinishedPulling="2025-10-02 09:58:59.253023053 +0000 UTC m=+1326.900708120" observedRunningTime="2025-10-02 09:59:00.72799019 +0000 UTC m=+1328.375675257" watchObservedRunningTime="2025-10-02 09:59:00.747902449 +0000 UTC m=+1328.395587516" Oct 02 09:59:02 crc kubenswrapper[4771]: I1002 09:59:02.362360 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-f206-account-create-nlvdg" Oct 02 09:59:02 crc kubenswrapper[4771]: I1002 09:59:02.461118 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk5lt\" (UniqueName: \"kubernetes.io/projected/55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc-kube-api-access-xk5lt\") pod \"55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc\" (UID: \"55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc\") " Oct 02 09:59:02 crc kubenswrapper[4771]: I1002 09:59:02.476103 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc-kube-api-access-xk5lt" (OuterVolumeSpecName: "kube-api-access-xk5lt") pod "55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc" (UID: "55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc"). InnerVolumeSpecName "kube-api-access-xk5lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:02 crc kubenswrapper[4771]: I1002 09:59:02.564432 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk5lt\" (UniqueName: \"kubernetes.io/projected/55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc-kube-api-access-xk5lt\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:02 crc kubenswrapper[4771]: I1002 09:59:02.703495 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-f206-account-create-nlvdg" event={"ID":"55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc","Type":"ContainerDied","Data":"e6ccd51ddad9f03a61f4694050b8661a57ee5f99376b641807fd92c7f37d3729"} Oct 02 09:59:02 crc kubenswrapper[4771]: I1002 09:59:02.703570 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6ccd51ddad9f03a61f4694050b8661a57ee5f99376b641807fd92c7f37d3729" Oct 02 09:59:02 crc kubenswrapper[4771]: I1002 09:59:02.703718 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-f206-account-create-nlvdg" Oct 02 09:59:03 crc kubenswrapper[4771]: I1002 09:59:03.426335 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:59:03 crc kubenswrapper[4771]: I1002 09:59:03.924377 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.779064 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Oct 02 09:59:04 crc kubenswrapper[4771]: E1002 09:59:04.779791 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc" containerName="mariadb-account-create" Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.779810 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc" containerName="mariadb-account-create" Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.780091 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc" containerName="mariadb-account-create" Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.780953 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.791926 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.831084 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.854654 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37a8032-f348-4e48-ac3a-9da319680ad2-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"e37a8032-f348-4e48-ac3a-9da319680ad2\") " pod="openstack/mysqld-exporter-0" Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.854838 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e37a8032-f348-4e48-ac3a-9da319680ad2-config-data\") pod \"mysqld-exporter-0\" (UID: \"e37a8032-f348-4e48-ac3a-9da319680ad2\") " pod="openstack/mysqld-exporter-0" Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.854904 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5cfh\" (UniqueName: \"kubernetes.io/projected/e37a8032-f348-4e48-ac3a-9da319680ad2-kube-api-access-s5cfh\") pod \"mysqld-exporter-0\" (UID: \"e37a8032-f348-4e48-ac3a-9da319680ad2\") " pod="openstack/mysqld-exporter-0" Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.957244 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e37a8032-f348-4e48-ac3a-9da319680ad2-config-data\") pod \"mysqld-exporter-0\" (UID: \"e37a8032-f348-4e48-ac3a-9da319680ad2\") " pod="openstack/mysqld-exporter-0" Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.957330 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5cfh\" (UniqueName: \"kubernetes.io/projected/e37a8032-f348-4e48-ac3a-9da319680ad2-kube-api-access-s5cfh\") pod \"mysqld-exporter-0\" (UID: \"e37a8032-f348-4e48-ac3a-9da319680ad2\") " pod="openstack/mysqld-exporter-0" Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.957516 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37a8032-f348-4e48-ac3a-9da319680ad2-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"e37a8032-f348-4e48-ac3a-9da319680ad2\") " pod="openstack/mysqld-exporter-0" Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.967992 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37a8032-f348-4e48-ac3a-9da319680ad2-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"e37a8032-f348-4e48-ac3a-9da319680ad2\") " pod="openstack/mysqld-exporter-0" Oct 02 09:59:04 crc kubenswrapper[4771]: I1002 09:59:04.969546 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e37a8032-f348-4e48-ac3a-9da319680ad2-config-data\") pod \"mysqld-exporter-0\" (UID: \"e37a8032-f348-4e48-ac3a-9da319680ad2\") " pod="openstack/mysqld-exporter-0" Oct 02 09:59:05 crc kubenswrapper[4771]: I1002 09:59:05.007837 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5cfh\" (UniqueName: \"kubernetes.io/projected/e37a8032-f348-4e48-ac3a-9da319680ad2-kube-api-access-s5cfh\") pod \"mysqld-exporter-0\" (UID: \"e37a8032-f348-4e48-ac3a-9da319680ad2\") " pod="openstack/mysqld-exporter-0" Oct 02 09:59:05 crc kubenswrapper[4771]: I1002 09:59:05.147714 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Oct 02 09:59:05 crc kubenswrapper[4771]: I1002 09:59:05.321035 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:05 crc kubenswrapper[4771]: I1002 09:59:05.787631 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Oct 02 09:59:05 crc kubenswrapper[4771]: W1002 09:59:05.842041 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode37a8032_f348_4e48_ac3a_9da319680ad2.slice/crio-d8e0fd6c12260c19548e5f5e34613f500735092853fc6e1cf9604725264ed6fe WatchSource:0}: Error finding container d8e0fd6c12260c19548e5f5e34613f500735092853fc6e1cf9604725264ed6fe: Status 404 returned error can't find the container with id d8e0fd6c12260c19548e5f5e34613f500735092853fc6e1cf9604725264ed6fe Oct 02 09:59:05 crc kubenswrapper[4771]: I1002 09:59:05.937516 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-74swj"] Oct 02 09:59:05 crc kubenswrapper[4771]: I1002 09:59:05.939764 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-74swj" Oct 02 09:59:05 crc kubenswrapper[4771]: I1002 09:59:05.956480 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-74swj"] Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.087415 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-5mhbt"] Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.089721 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5mhbt" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.100222 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-5mhbt"] Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.107210 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp6sz\" (UniqueName: \"kubernetes.io/projected/099832fe-d356-4281-a374-127ce62a5fee-kube-api-access-tp6sz\") pod \"heat-db-create-74swj\" (UID: \"099832fe-d356-4281-a374-127ce62a5fee\") " pod="openstack/heat-db-create-74swj" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.191037 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-m55ds"] Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.193037 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m55ds" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.209902 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp6sz\" (UniqueName: \"kubernetes.io/projected/099832fe-d356-4281-a374-127ce62a5fee-kube-api-access-tp6sz\") pod \"heat-db-create-74swj\" (UID: \"099832fe-d356-4281-a374-127ce62a5fee\") " pod="openstack/heat-db-create-74swj" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.213855 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-m55ds"] Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.214143 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djzdb\" (UniqueName: \"kubernetes.io/projected/4a3923b9-976f-46e8-a672-b56698a83cbb-kube-api-access-djzdb\") pod \"cinder-db-create-5mhbt\" (UID: \"4a3923b9-976f-46e8-a672-b56698a83cbb\") " pod="openstack/cinder-db-create-5mhbt" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.284755 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp6sz\" (UniqueName: \"kubernetes.io/projected/099832fe-d356-4281-a374-127ce62a5fee-kube-api-access-tp6sz\") pod \"heat-db-create-74swj\" (UID: \"099832fe-d356-4281-a374-127ce62a5fee\") " pod="openstack/heat-db-create-74swj" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.316939 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp4vj\" (UniqueName: \"kubernetes.io/projected/709fe2b2-e133-4651-8d6a-ff41bb580e23-kube-api-access-pp4vj\") pod \"barbican-db-create-m55ds\" (UID: \"709fe2b2-e133-4651-8d6a-ff41bb580e23\") " pod="openstack/barbican-db-create-m55ds" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.317292 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djzdb\" (UniqueName: \"kubernetes.io/projected/4a3923b9-976f-46e8-a672-b56698a83cbb-kube-api-access-djzdb\") pod \"cinder-db-create-5mhbt\" (UID: \"4a3923b9-976f-46e8-a672-b56698a83cbb\") " pod="openstack/cinder-db-create-5mhbt" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.348068 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djzdb\" (UniqueName: \"kubernetes.io/projected/4a3923b9-976f-46e8-a672-b56698a83cbb-kube-api-access-djzdb\") pod \"cinder-db-create-5mhbt\" (UID: \"4a3923b9-976f-46e8-a672-b56698a83cbb\") " pod="openstack/cinder-db-create-5mhbt" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.395227 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-krvcs"] Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.400608 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-krvcs" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.408682 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5mhbt" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.412148 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-krvcs"] Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.419572 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp4vj\" (UniqueName: \"kubernetes.io/projected/709fe2b2-e133-4651-8d6a-ff41bb580e23-kube-api-access-pp4vj\") pod \"barbican-db-create-m55ds\" (UID: \"709fe2b2-e133-4651-8d6a-ff41bb580e23\") " pod="openstack/barbican-db-create-m55ds" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.451756 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp4vj\" (UniqueName: \"kubernetes.io/projected/709fe2b2-e133-4651-8d6a-ff41bb580e23-kube-api-access-pp4vj\") pod \"barbican-db-create-m55ds\" (UID: \"709fe2b2-e133-4651-8d6a-ff41bb580e23\") " pod="openstack/barbican-db-create-m55ds" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.516588 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-cjqrx"] Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.518875 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cjqrx" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.522478 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.522840 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.523979 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp897\" (UniqueName: \"kubernetes.io/projected/64ee0cf5-2677-4bc4-bf6d-ae15055a2689-kube-api-access-xp897\") pod \"neutron-db-create-krvcs\" (UID: \"64ee0cf5-2677-4bc4-bf6d-ae15055a2689\") " pod="openstack/neutron-db-create-krvcs" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.526596 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8kg7s" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.526748 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.534629 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m55ds" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.568378 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-cjqrx"] Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.570739 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-74swj" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.626537 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-combined-ca-bundle\") pod \"keystone-db-sync-cjqrx\" (UID: \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\") " pod="openstack/keystone-db-sync-cjqrx" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.626619 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-config-data\") pod \"keystone-db-sync-cjqrx\" (UID: \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\") " pod="openstack/keystone-db-sync-cjqrx" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.629063 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbfbp\" (UniqueName: \"kubernetes.io/projected/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-kube-api-access-mbfbp\") pod \"keystone-db-sync-cjqrx\" (UID: \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\") " pod="openstack/keystone-db-sync-cjqrx" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.629256 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp897\" (UniqueName: \"kubernetes.io/projected/64ee0cf5-2677-4bc4-bf6d-ae15055a2689-kube-api-access-xp897\") pod \"neutron-db-create-krvcs\" (UID: \"64ee0cf5-2677-4bc4-bf6d-ae15055a2689\") " pod="openstack/neutron-db-create-krvcs" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.663775 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp897\" (UniqueName: \"kubernetes.io/projected/64ee0cf5-2677-4bc4-bf6d-ae15055a2689-kube-api-access-xp897\") pod \"neutron-db-create-krvcs\" (UID: \"64ee0cf5-2677-4bc4-bf6d-ae15055a2689\") " pod="openstack/neutron-db-create-krvcs" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.726833 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-krvcs" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.742082 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-combined-ca-bundle\") pod \"keystone-db-sync-cjqrx\" (UID: \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\") " pod="openstack/keystone-db-sync-cjqrx" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.742774 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-config-data\") pod \"keystone-db-sync-cjqrx\" (UID: \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\") " pod="openstack/keystone-db-sync-cjqrx" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.742977 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbfbp\" (UniqueName: \"kubernetes.io/projected/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-kube-api-access-mbfbp\") pod \"keystone-db-sync-cjqrx\" (UID: \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\") " pod="openstack/keystone-db-sync-cjqrx" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.755674 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-combined-ca-bundle\") pod \"keystone-db-sync-cjqrx\" (UID: \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\") " pod="openstack/keystone-db-sync-cjqrx" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.762668 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-config-data\") pod \"keystone-db-sync-cjqrx\" (UID: \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\") " pod="openstack/keystone-db-sync-cjqrx" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.783173 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbfbp\" (UniqueName: \"kubernetes.io/projected/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-kube-api-access-mbfbp\") pod \"keystone-db-sync-cjqrx\" (UID: \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\") " pod="openstack/keystone-db-sync-cjqrx" Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.851278 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"e37a8032-f348-4e48-ac3a-9da319680ad2","Type":"ContainerStarted","Data":"d8e0fd6c12260c19548e5f5e34613f500735092853fc6e1cf9604725264ed6fe"} Oct 02 09:59:06 crc kubenswrapper[4771]: I1002 09:59:06.894798 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cjqrx" Oct 02 09:59:07 crc kubenswrapper[4771]: I1002 09:59:07.280402 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-5mhbt"] Oct 02 09:59:07 crc kubenswrapper[4771]: I1002 09:59:07.395784 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-74swj"] Oct 02 09:59:07 crc kubenswrapper[4771]: I1002 09:59:07.627433 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-m55ds"] Oct 02 09:59:07 crc kubenswrapper[4771]: I1002 09:59:07.645208 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-krvcs"] Oct 02 09:59:07 crc kubenswrapper[4771]: I1002 09:59:07.864117 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-cjqrx"] Oct 02 09:59:08 crc kubenswrapper[4771]: W1002 09:59:08.360304 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64ee0cf5_2677_4bc4_bf6d_ae15055a2689.slice/crio-7bf1064171e9d1ea100478cd1d7a20e9958c133d0228f928274636ce3a8f8e08 WatchSource:0}: Error finding container 7bf1064171e9d1ea100478cd1d7a20e9958c133d0228f928274636ce3a8f8e08: Status 404 returned error can't find the container with id 7bf1064171e9d1ea100478cd1d7a20e9958c133d0228f928274636ce3a8f8e08 Oct 02 09:59:08 crc kubenswrapper[4771]: I1002 09:59:08.885163 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-74swj" event={"ID":"099832fe-d356-4281-a374-127ce62a5fee","Type":"ContainerStarted","Data":"a4985f6bcd078d5b7bf4e8e8b365d58b6b3a3b4650c42b20bb059a2c34866a52"} Oct 02 09:59:08 crc kubenswrapper[4771]: I1002 09:59:08.888199 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-krvcs" event={"ID":"64ee0cf5-2677-4bc4-bf6d-ae15055a2689","Type":"ContainerStarted","Data":"7bf1064171e9d1ea100478cd1d7a20e9958c133d0228f928274636ce3a8f8e08"} Oct 02 09:59:08 crc kubenswrapper[4771]: I1002 09:59:08.890419 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m55ds" event={"ID":"709fe2b2-e133-4651-8d6a-ff41bb580e23","Type":"ContainerStarted","Data":"729306b1809bbcc9904dbe41f3390f0ad04f791df9663cdf0b09d696fe36a8c7"} Oct 02 09:59:08 crc kubenswrapper[4771]: I1002 09:59:08.891708 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cjqrx" event={"ID":"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a","Type":"ContainerStarted","Data":"fc3175fb2bc3c0d7ba4850ec237eccf26391a2694749b11739348a4436236e84"} Oct 02 09:59:09 crc kubenswrapper[4771]: I1002 09:59:09.915264 4771 generic.go:334] "Generic (PLEG): container finished" podID="087b7d39-d0cf-4896-8220-d79e94810ff5" containerID="44a564538fad32d29e4ba361673681ee694bfa50ed1594361eaa7d7e84a46bf3" exitCode=0 Oct 02 09:59:09 crc kubenswrapper[4771]: I1002 09:59:09.915443 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-c9bwn" event={"ID":"087b7d39-d0cf-4896-8220-d79e94810ff5","Type":"ContainerDied","Data":"44a564538fad32d29e4ba361673681ee694bfa50ed1594361eaa7d7e84a46bf3"} Oct 02 09:59:11 crc kubenswrapper[4771]: I1002 09:59:11.889230 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:59:11 crc kubenswrapper[4771]: I1002 09:59:11.899333 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4c4deacb-7b2f-49d1-8c7e-b87913f027d1-etc-swift\") pod \"swift-storage-0\" (UID: \"4c4deacb-7b2f-49d1-8c7e-b87913f027d1\") " pod="openstack/swift-storage-0" Oct 02 09:59:12 crc kubenswrapper[4771]: I1002 09:59:12.146038 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:59:12 crc kubenswrapper[4771]: I1002 09:59:12.146222 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:59:12 crc kubenswrapper[4771]: I1002 09:59:12.162365 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 09:59:15 crc kubenswrapper[4771]: I1002 09:59:15.320619 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:15 crc kubenswrapper[4771]: I1002 09:59:15.325920 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.003571 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-c9bwn" event={"ID":"087b7d39-d0cf-4896-8220-d79e94810ff5","Type":"ContainerDied","Data":"5722c920dc00e8b6342fa579faccc501881e9f5a0edb9545ce3c4e5eed845850"} Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.003948 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5722c920dc00e8b6342fa579faccc501881e9f5a0edb9545ce3c4e5eed845850" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.009700 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5mhbt" event={"ID":"4a3923b9-976f-46e8-a672-b56698a83cbb","Type":"ContainerStarted","Data":"cb439e2f3ebc13c2f510354a305a156747b4d5d21f70c6a874ea30b268c5648f"} Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.011396 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.254542 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.409587 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f66lq\" (UniqueName: \"kubernetes.io/projected/087b7d39-d0cf-4896-8220-d79e94810ff5-kube-api-access-f66lq\") pod \"087b7d39-d0cf-4896-8220-d79e94810ff5\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.409753 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-swiftconf\") pod \"087b7d39-d0cf-4896-8220-d79e94810ff5\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.409809 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/087b7d39-d0cf-4896-8220-d79e94810ff5-scripts\") pod \"087b7d39-d0cf-4896-8220-d79e94810ff5\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.409843 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-combined-ca-bundle\") pod \"087b7d39-d0cf-4896-8220-d79e94810ff5\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.410090 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-dispersionconf\") pod \"087b7d39-d0cf-4896-8220-d79e94810ff5\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.410154 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/087b7d39-d0cf-4896-8220-d79e94810ff5-ring-data-devices\") pod \"087b7d39-d0cf-4896-8220-d79e94810ff5\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.410226 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/087b7d39-d0cf-4896-8220-d79e94810ff5-etc-swift\") pod \"087b7d39-d0cf-4896-8220-d79e94810ff5\" (UID: \"087b7d39-d0cf-4896-8220-d79e94810ff5\") " Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.412780 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/087b7d39-d0cf-4896-8220-d79e94810ff5-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "087b7d39-d0cf-4896-8220-d79e94810ff5" (UID: "087b7d39-d0cf-4896-8220-d79e94810ff5"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.413997 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/087b7d39-d0cf-4896-8220-d79e94810ff5-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "087b7d39-d0cf-4896-8220-d79e94810ff5" (UID: "087b7d39-d0cf-4896-8220-d79e94810ff5"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.421230 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/087b7d39-d0cf-4896-8220-d79e94810ff5-kube-api-access-f66lq" (OuterVolumeSpecName: "kube-api-access-f66lq") pod "087b7d39-d0cf-4896-8220-d79e94810ff5" (UID: "087b7d39-d0cf-4896-8220-d79e94810ff5"). InnerVolumeSpecName "kube-api-access-f66lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.443965 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "087b7d39-d0cf-4896-8220-d79e94810ff5" (UID: "087b7d39-d0cf-4896-8220-d79e94810ff5"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.515307 4771 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/087b7d39-d0cf-4896-8220-d79e94810ff5-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.515343 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f66lq\" (UniqueName: \"kubernetes.io/projected/087b7d39-d0cf-4896-8220-d79e94810ff5-kube-api-access-f66lq\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.515359 4771 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.515371 4771 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/087b7d39-d0cf-4896-8220-d79e94810ff5-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.593055 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "087b7d39-d0cf-4896-8220-d79e94810ff5" (UID: "087b7d39-d0cf-4896-8220-d79e94810ff5"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.598838 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "087b7d39-d0cf-4896-8220-d79e94810ff5" (UID: "087b7d39-d0cf-4896-8220-d79e94810ff5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.620643 4771 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.620978 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087b7d39-d0cf-4896-8220-d79e94810ff5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.626866 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/087b7d39-d0cf-4896-8220-d79e94810ff5-scripts" (OuterVolumeSpecName: "scripts") pod "087b7d39-d0cf-4896-8220-d79e94810ff5" (UID: "087b7d39-d0cf-4896-8220-d79e94810ff5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.662764 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 09:59:16 crc kubenswrapper[4771]: I1002 09:59:16.724801 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/087b7d39-d0cf-4896-8220-d79e94810ff5-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.023480 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"12580071e0584b75ebe269e22eae132c45c45ab71c0338ca0d4f39f5b7ac60dd"} Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.025618 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-t42dx" event={"ID":"4c6203ef-b851-49ba-ae59-9b62c473c497","Type":"ContainerStarted","Data":"d66e976c25ca5a30398fd5404e8f0a51209b789267b8e1d2ab18533250acb02f"} Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.027410 4771 generic.go:334] "Generic (PLEG): container finished" podID="64ee0cf5-2677-4bc4-bf6d-ae15055a2689" containerID="d89bdd71c6f86666ce173a64b52637a86ceb3e85900ca707502c6fb2922f51a8" exitCode=0 Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.027450 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-krvcs" event={"ID":"64ee0cf5-2677-4bc4-bf6d-ae15055a2689","Type":"ContainerDied","Data":"d89bdd71c6f86666ce173a64b52637a86ceb3e85900ca707502c6fb2922f51a8"} Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.029025 4771 generic.go:334] "Generic (PLEG): container finished" podID="709fe2b2-e133-4651-8d6a-ff41bb580e23" containerID="ec922be9092fad9973c5343b20ed6675d419ee21e23d69688a1527803a5a240b" exitCode=0 Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.029066 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m55ds" event={"ID":"709fe2b2-e133-4651-8d6a-ff41bb580e23","Type":"ContainerDied","Data":"ec922be9092fad9973c5343b20ed6675d419ee21e23d69688a1527803a5a240b"} Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.030643 4771 generic.go:334] "Generic (PLEG): container finished" podID="099832fe-d356-4281-a374-127ce62a5fee" containerID="3c41f352c6503f3902610c6f2fc0bbf3c2b1eb173b36d6c2eeae980c97fea62d" exitCode=0 Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.030687 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-74swj" event={"ID":"099832fe-d356-4281-a374-127ce62a5fee","Type":"ContainerDied","Data":"3c41f352c6503f3902610c6f2fc0bbf3c2b1eb173b36d6c2eeae980c97fea62d"} Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.033744 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"e37a8032-f348-4e48-ac3a-9da319680ad2","Type":"ContainerStarted","Data":"c0e9c71d39ee3db61f689d3ee4d53ba4357ca3415bb3eb88f0392121dbb5f8b6"} Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.035585 4771 generic.go:334] "Generic (PLEG): container finished" podID="4a3923b9-976f-46e8-a672-b56698a83cbb" containerID="db924463bbf11fd168d26cff79a6e23cbd8dc5b01db6797c5cb332d55df9a30e" exitCode=0 Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.035663 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5mhbt" event={"ID":"4a3923b9-976f-46e8-a672-b56698a83cbb","Type":"ContainerDied","Data":"db924463bbf11fd168d26cff79a6e23cbd8dc5b01db6797c5cb332d55df9a30e"} Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.035801 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-c9bwn" Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.043049 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-t42dx" podStartSLOduration=8.222730049 podStartE2EDuration="25.043022438s" podCreationTimestamp="2025-10-02 09:58:52 +0000 UTC" firstStartedPulling="2025-10-02 09:58:59.258242257 +0000 UTC m=+1326.905927324" lastFinishedPulling="2025-10-02 09:59:16.078534656 +0000 UTC m=+1343.726219713" observedRunningTime="2025-10-02 09:59:17.042699519 +0000 UTC m=+1344.690384586" watchObservedRunningTime="2025-10-02 09:59:17.043022438 +0000 UTC m=+1344.690707515" Oct 02 09:59:17 crc kubenswrapper[4771]: I1002 09:59:17.146797 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=3.087760693 podStartE2EDuration="13.146778221s" podCreationTimestamp="2025-10-02 09:59:04 +0000 UTC" firstStartedPulling="2025-10-02 09:59:05.88745032 +0000 UTC m=+1333.535135388" lastFinishedPulling="2025-10-02 09:59:15.946467849 +0000 UTC m=+1343.594152916" observedRunningTime="2025-10-02 09:59:17.124997364 +0000 UTC m=+1344.772682451" watchObservedRunningTime="2025-10-02 09:59:17.146778221 +0000 UTC m=+1344.794463278" Oct 02 09:59:18 crc kubenswrapper[4771]: I1002 09:59:18.720570 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 09:59:18 crc kubenswrapper[4771]: I1002 09:59:18.721457 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="prometheus" containerID="cri-o://c3286830dd0529949a1d53aeff88ff2c8e1154091dff38d555badaf1e6af5d75" gracePeriod=600 Oct 02 09:59:18 crc kubenswrapper[4771]: I1002 09:59:18.722044 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="thanos-sidecar" containerID="cri-o://61c3b16f18cbb70283e079c3cf0a250091cf91ceb3e801c047cf6918fb937be1" gracePeriod=600 Oct 02 09:59:18 crc kubenswrapper[4771]: I1002 09:59:18.722111 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="config-reloader" containerID="cri-o://7e88704fb0629cfd87c9422d98c45c8ecb8684b1aa2ddf1d73efee5edce3e523" gracePeriod=600 Oct 02 09:59:18 crc kubenswrapper[4771]: I1002 09:59:18.979435 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-krvcs" Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.088485 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xp897\" (UniqueName: \"kubernetes.io/projected/64ee0cf5-2677-4bc4-bf6d-ae15055a2689-kube-api-access-xp897\") pod \"64ee0cf5-2677-4bc4-bf6d-ae15055a2689\" (UID: \"64ee0cf5-2677-4bc4-bf6d-ae15055a2689\") " Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.099527 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64ee0cf5-2677-4bc4-bf6d-ae15055a2689-kube-api-access-xp897" (OuterVolumeSpecName: "kube-api-access-xp897") pod "64ee0cf5-2677-4bc4-bf6d-ae15055a2689" (UID: "64ee0cf5-2677-4bc4-bf6d-ae15055a2689"). InnerVolumeSpecName "kube-api-access-xp897". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.167012 4771 generic.go:334] "Generic (PLEG): container finished" podID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerID="61c3b16f18cbb70283e079c3cf0a250091cf91ceb3e801c047cf6918fb937be1" exitCode=0 Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.167052 4771 generic.go:334] "Generic (PLEG): container finished" podID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerID="c3286830dd0529949a1d53aeff88ff2c8e1154091dff38d555badaf1e6af5d75" exitCode=0 Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.170513 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"95cd3d5a-1000-4212-b4d9-5d0da549c7f3","Type":"ContainerDied","Data":"61c3b16f18cbb70283e079c3cf0a250091cf91ceb3e801c047cf6918fb937be1"} Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.170564 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"95cd3d5a-1000-4212-b4d9-5d0da549c7f3","Type":"ContainerDied","Data":"c3286830dd0529949a1d53aeff88ff2c8e1154091dff38d555badaf1e6af5d75"} Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.199018 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xp897\" (UniqueName: \"kubernetes.io/projected/64ee0cf5-2677-4bc4-bf6d-ae15055a2689-kube-api-access-xp897\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.226417 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-krvcs" event={"ID":"64ee0cf5-2677-4bc4-bf6d-ae15055a2689","Type":"ContainerDied","Data":"7bf1064171e9d1ea100478cd1d7a20e9958c133d0228f928274636ce3a8f8e08"} Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.226689 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bf1064171e9d1ea100478cd1d7a20e9958c133d0228f928274636ce3a8f8e08" Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.226823 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-krvcs" Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.555411 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5mhbt" Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.633500 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djzdb\" (UniqueName: \"kubernetes.io/projected/4a3923b9-976f-46e8-a672-b56698a83cbb-kube-api-access-djzdb\") pod \"4a3923b9-976f-46e8-a672-b56698a83cbb\" (UID: \"4a3923b9-976f-46e8-a672-b56698a83cbb\") " Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.645523 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a3923b9-976f-46e8-a672-b56698a83cbb-kube-api-access-djzdb" (OuterVolumeSpecName: "kube-api-access-djzdb") pod "4a3923b9-976f-46e8-a672-b56698a83cbb" (UID: "4a3923b9-976f-46e8-a672-b56698a83cbb"). InnerVolumeSpecName "kube-api-access-djzdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.737167 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djzdb\" (UniqueName: \"kubernetes.io/projected/4a3923b9-976f-46e8-a672-b56698a83cbb-kube-api-access-djzdb\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.738334 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-74swj" Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.843002 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp6sz\" (UniqueName: \"kubernetes.io/projected/099832fe-d356-4281-a374-127ce62a5fee-kube-api-access-tp6sz\") pod \"099832fe-d356-4281-a374-127ce62a5fee\" (UID: \"099832fe-d356-4281-a374-127ce62a5fee\") " Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.855935 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/099832fe-d356-4281-a374-127ce62a5fee-kube-api-access-tp6sz" (OuterVolumeSpecName: "kube-api-access-tp6sz") pod "099832fe-d356-4281-a374-127ce62a5fee" (UID: "099832fe-d356-4281-a374-127ce62a5fee"). InnerVolumeSpecName "kube-api-access-tp6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:19 crc kubenswrapper[4771]: I1002 09:59:19.947971 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp6sz\" (UniqueName: \"kubernetes.io/projected/099832fe-d356-4281-a374-127ce62a5fee-kube-api-access-tp6sz\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.092450 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m55ds" Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.164627 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pp4vj\" (UniqueName: \"kubernetes.io/projected/709fe2b2-e133-4651-8d6a-ff41bb580e23-kube-api-access-pp4vj\") pod \"709fe2b2-e133-4651-8d6a-ff41bb580e23\" (UID: \"709fe2b2-e133-4651-8d6a-ff41bb580e23\") " Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.206360 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/709fe2b2-e133-4651-8d6a-ff41bb580e23-kube-api-access-pp4vj" (OuterVolumeSpecName: "kube-api-access-pp4vj") pod "709fe2b2-e133-4651-8d6a-ff41bb580e23" (UID: "709fe2b2-e133-4651-8d6a-ff41bb580e23"). InnerVolumeSpecName "kube-api-access-pp4vj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.273001 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pp4vj\" (UniqueName: \"kubernetes.io/projected/709fe2b2-e133-4651-8d6a-ff41bb580e23-kube-api-access-pp4vj\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.280640 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-74swj" event={"ID":"099832fe-d356-4281-a374-127ce62a5fee","Type":"ContainerDied","Data":"a4985f6bcd078d5b7bf4e8e8b365d58b6b3a3b4650c42b20bb059a2c34866a52"} Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.280688 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4985f6bcd078d5b7bf4e8e8b365d58b6b3a3b4650c42b20bb059a2c34866a52" Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.280695 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-74swj" Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.283623 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5mhbt" event={"ID":"4a3923b9-976f-46e8-a672-b56698a83cbb","Type":"ContainerDied","Data":"cb439e2f3ebc13c2f510354a305a156747b4d5d21f70c6a874ea30b268c5648f"} Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.283650 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb439e2f3ebc13c2f510354a305a156747b4d5d21f70c6a874ea30b268c5648f" Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.283717 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5mhbt" Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.291453 4771 generic.go:334] "Generic (PLEG): container finished" podID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerID="7e88704fb0629cfd87c9422d98c45c8ecb8684b1aa2ddf1d73efee5edce3e523" exitCode=0 Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.291542 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"95cd3d5a-1000-4212-b4d9-5d0da549c7f3","Type":"ContainerDied","Data":"7e88704fb0629cfd87c9422d98c45c8ecb8684b1aa2ddf1d73efee5edce3e523"} Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.296944 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"d2845f97dbfcea0d981fb200cfe56bc919fdbe98166bed572af65fe299e7da16"} Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.302968 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m55ds" event={"ID":"709fe2b2-e133-4651-8d6a-ff41bb580e23","Type":"ContainerDied","Data":"729306b1809bbcc9904dbe41f3390f0ad04f791df9663cdf0b09d696fe36a8c7"} Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.303009 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="729306b1809bbcc9904dbe41f3390f0ad04f791df9663cdf0b09d696fe36a8c7" Oct 02 09:59:20 crc kubenswrapper[4771]: I1002 09:59:20.303069 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m55ds" Oct 02 09:59:21 crc kubenswrapper[4771]: I1002 09:59:21.318959 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"c599e7746a3e4c7ed93b6c5d628d68929a17a58d89f8b6d2f933774bdc7fbc81"} Oct 02 09:59:21 crc kubenswrapper[4771]: I1002 09:59:21.729714 4771 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod7eb14f25-cd51-4f54-ac31-3eb9cf51e152"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod7eb14f25-cd51-4f54-ac31-3eb9cf51e152] : Timed out while waiting for systemd to remove kubepods-besteffort-pod7eb14f25_cd51_4f54_ac31_3eb9cf51e152.slice" Oct 02 09:59:21 crc kubenswrapper[4771]: E1002 09:59:21.729764 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod7eb14f25-cd51-4f54-ac31-3eb9cf51e152] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod7eb14f25-cd51-4f54-ac31-3eb9cf51e152] : Timed out while waiting for systemd to remove kubepods-besteffort-pod7eb14f25_cd51_4f54_ac31_3eb9cf51e152.slice" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" podUID="7eb14f25-cd51-4f54-ac31-3eb9cf51e152" Oct 02 09:59:22 crc kubenswrapper[4771]: I1002 09:59:22.331536 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-cwm9n" Oct 02 09:59:22 crc kubenswrapper[4771]: I1002 09:59:22.378743 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-cwm9n"] Oct 02 09:59:22 crc kubenswrapper[4771]: I1002 09:59:22.390334 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-cwm9n"] Oct 02 09:59:23 crc kubenswrapper[4771]: I1002 09:59:23.321908 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.138:9090/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 09:59:23 crc kubenswrapper[4771]: I1002 09:59:23.701621 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7eb14f25-cd51-4f54-ac31-3eb9cf51e152" path="/var/lib/kubelet/pods/7eb14f25-cd51-4f54-ac31-3eb9cf51e152/volumes" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.068099 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-73f2-account-create-phct2"] Oct 02 09:59:26 crc kubenswrapper[4771]: E1002 09:59:26.071689 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a3923b9-976f-46e8-a672-b56698a83cbb" containerName="mariadb-database-create" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.071749 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a3923b9-976f-46e8-a672-b56698a83cbb" containerName="mariadb-database-create" Oct 02 09:59:26 crc kubenswrapper[4771]: E1002 09:59:26.071785 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64ee0cf5-2677-4bc4-bf6d-ae15055a2689" containerName="mariadb-database-create" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.071795 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="64ee0cf5-2677-4bc4-bf6d-ae15055a2689" containerName="mariadb-database-create" Oct 02 09:59:26 crc kubenswrapper[4771]: E1002 09:59:26.071817 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087b7d39-d0cf-4896-8220-d79e94810ff5" containerName="swift-ring-rebalance" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.071824 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="087b7d39-d0cf-4896-8220-d79e94810ff5" containerName="swift-ring-rebalance" Oct 02 09:59:26 crc kubenswrapper[4771]: E1002 09:59:26.071842 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="099832fe-d356-4281-a374-127ce62a5fee" containerName="mariadb-database-create" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.071849 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="099832fe-d356-4281-a374-127ce62a5fee" containerName="mariadb-database-create" Oct 02 09:59:26 crc kubenswrapper[4771]: E1002 09:59:26.071875 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="709fe2b2-e133-4651-8d6a-ff41bb580e23" containerName="mariadb-database-create" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.071882 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="709fe2b2-e133-4651-8d6a-ff41bb580e23" containerName="mariadb-database-create" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.072321 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="099832fe-d356-4281-a374-127ce62a5fee" containerName="mariadb-database-create" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.072341 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="64ee0cf5-2677-4bc4-bf6d-ae15055a2689" containerName="mariadb-database-create" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.072355 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="709fe2b2-e133-4651-8d6a-ff41bb580e23" containerName="mariadb-database-create" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.072373 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a3923b9-976f-46e8-a672-b56698a83cbb" containerName="mariadb-database-create" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.072392 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="087b7d39-d0cf-4896-8220-d79e94810ff5" containerName="swift-ring-rebalance" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.073460 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-73f2-account-create-phct2" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.077256 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.084658 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-73f2-account-create-phct2"] Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.225922 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh8fb\" (UniqueName: \"kubernetes.io/projected/b63a764e-b703-434d-860d-7774c43035aa-kube-api-access-xh8fb\") pod \"cinder-73f2-account-create-phct2\" (UID: \"b63a764e-b703-434d-860d-7774c43035aa\") " pod="openstack/cinder-73f2-account-create-phct2" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.328873 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh8fb\" (UniqueName: \"kubernetes.io/projected/b63a764e-b703-434d-860d-7774c43035aa-kube-api-access-xh8fb\") pod \"cinder-73f2-account-create-phct2\" (UID: \"b63a764e-b703-434d-860d-7774c43035aa\") " pod="openstack/cinder-73f2-account-create-phct2" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.358431 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-4b06-account-create-5s4cn"] Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.360345 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-4b06-account-create-5s4cn" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.363480 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.369822 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-4b06-account-create-5s4cn"] Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.385359 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh8fb\" (UniqueName: \"kubernetes.io/projected/b63a764e-b703-434d-860d-7774c43035aa-kube-api-access-xh8fb\") pod \"cinder-73f2-account-create-phct2\" (UID: \"b63a764e-b703-434d-860d-7774c43035aa\") " pod="openstack/cinder-73f2-account-create-phct2" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.431494 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-73f2-account-create-phct2" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.516837 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-816a-account-create-6pk4g"] Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.526112 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-816a-account-create-6pk4g" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.530724 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.531107 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-816a-account-create-6pk4g"] Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.541353 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4mhb\" (UniqueName: \"kubernetes.io/projected/fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d-kube-api-access-m4mhb\") pod \"heat-4b06-account-create-5s4cn\" (UID: \"fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d\") " pod="openstack/heat-4b06-account-create-5s4cn" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.644183 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4mhb\" (UniqueName: \"kubernetes.io/projected/fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d-kube-api-access-m4mhb\") pod \"heat-4b06-account-create-5s4cn\" (UID: \"fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d\") " pod="openstack/heat-4b06-account-create-5s4cn" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.644336 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdc94\" (UniqueName: \"kubernetes.io/projected/6cc9f109-903c-4e52-8031-56898f9128ad-kube-api-access-qdc94\") pod \"barbican-816a-account-create-6pk4g\" (UID: \"6cc9f109-903c-4e52-8031-56898f9128ad\") " pod="openstack/barbican-816a-account-create-6pk4g" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.665950 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4mhb\" (UniqueName: \"kubernetes.io/projected/fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d-kube-api-access-m4mhb\") pod \"heat-4b06-account-create-5s4cn\" (UID: \"fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d\") " pod="openstack/heat-4b06-account-create-5s4cn" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.666460 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-9938-account-create-jj42p"] Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.668559 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9938-account-create-jj42p" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.671043 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.683812 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9938-account-create-jj42p"] Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.746614 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdc94\" (UniqueName: \"kubernetes.io/projected/6cc9f109-903c-4e52-8031-56898f9128ad-kube-api-access-qdc94\") pod \"barbican-816a-account-create-6pk4g\" (UID: \"6cc9f109-903c-4e52-8031-56898f9128ad\") " pod="openstack/barbican-816a-account-create-6pk4g" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.747107 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf2ms\" (UniqueName: \"kubernetes.io/projected/e215bc59-d489-4b95-8348-fe23e36f2d6e-kube-api-access-bf2ms\") pod \"neutron-9938-account-create-jj42p\" (UID: \"e215bc59-d489-4b95-8348-fe23e36f2d6e\") " pod="openstack/neutron-9938-account-create-jj42p" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.769783 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdc94\" (UniqueName: \"kubernetes.io/projected/6cc9f109-903c-4e52-8031-56898f9128ad-kube-api-access-qdc94\") pod \"barbican-816a-account-create-6pk4g\" (UID: \"6cc9f109-903c-4e52-8031-56898f9128ad\") " pod="openstack/barbican-816a-account-create-6pk4g" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.770412 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-4b06-account-create-5s4cn" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.849354 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf2ms\" (UniqueName: \"kubernetes.io/projected/e215bc59-d489-4b95-8348-fe23e36f2d6e-kube-api-access-bf2ms\") pod \"neutron-9938-account-create-jj42p\" (UID: \"e215bc59-d489-4b95-8348-fe23e36f2d6e\") " pod="openstack/neutron-9938-account-create-jj42p" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.867521 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-816a-account-create-6pk4g" Oct 02 09:59:26 crc kubenswrapper[4771]: I1002 09:59:26.869565 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf2ms\" (UniqueName: \"kubernetes.io/projected/e215bc59-d489-4b95-8348-fe23e36f2d6e-kube-api-access-bf2ms\") pod \"neutron-9938-account-create-jj42p\" (UID: \"e215bc59-d489-4b95-8348-fe23e36f2d6e\") " pod="openstack/neutron-9938-account-create-jj42p" Oct 02 09:59:27 crc kubenswrapper[4771]: I1002 09:59:27.042536 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9938-account-create-jj42p" Oct 02 09:59:28 crc kubenswrapper[4771]: I1002 09:59:28.320999 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.138:9090/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 09:59:29 crc kubenswrapper[4771]: I1002 09:59:29.913372 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.026969 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-tls-assets\") pod \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.027020 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-web-config\") pod \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.027159 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-config-out\") pod \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.027289 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-thanos-prometheus-http-client-file\") pod \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.027503 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f33acc08-4fd4-473f-9263-ec4841f48809\") pod \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.027567 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lknt7\" (UniqueName: \"kubernetes.io/projected/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-kube-api-access-lknt7\") pod \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.027601 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-config\") pod \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.027628 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-prometheus-metric-storage-rulefiles-0\") pod \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\" (UID: \"95cd3d5a-1000-4212-b4d9-5d0da549c7f3\") " Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.029084 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "95cd3d5a-1000-4212-b4d9-5d0da549c7f3" (UID: "95cd3d5a-1000-4212-b4d9-5d0da549c7f3"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.056316 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-config-out" (OuterVolumeSpecName: "config-out") pod "95cd3d5a-1000-4212-b4d9-5d0da549c7f3" (UID: "95cd3d5a-1000-4212-b4d9-5d0da549c7f3"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.062519 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-config" (OuterVolumeSpecName: "config") pod "95cd3d5a-1000-4212-b4d9-5d0da549c7f3" (UID: "95cd3d5a-1000-4212-b4d9-5d0da549c7f3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.063514 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-kube-api-access-lknt7" (OuterVolumeSpecName: "kube-api-access-lknt7") pod "95cd3d5a-1000-4212-b4d9-5d0da549c7f3" (UID: "95cd3d5a-1000-4212-b4d9-5d0da549c7f3"). InnerVolumeSpecName "kube-api-access-lknt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.065394 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "95cd3d5a-1000-4212-b4d9-5d0da549c7f3" (UID: "95cd3d5a-1000-4212-b4d9-5d0da549c7f3"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.073418 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "95cd3d5a-1000-4212-b4d9-5d0da549c7f3" (UID: "95cd3d5a-1000-4212-b4d9-5d0da549c7f3"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.131547 4771 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.131586 4771 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-config-out\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.131599 4771 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.131614 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lknt7\" (UniqueName: \"kubernetes.io/projected/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-kube-api-access-lknt7\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.131625 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.131636 4771 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.190295 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-web-config" (OuterVolumeSpecName: "web-config") pod "95cd3d5a-1000-4212-b4d9-5d0da549c7f3" (UID: "95cd3d5a-1000-4212-b4d9-5d0da549c7f3"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.235912 4771 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/95cd3d5a-1000-4212-b4d9-5d0da549c7f3-web-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.245427 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f33acc08-4fd4-473f-9263-ec4841f48809" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "95cd3d5a-1000-4212-b4d9-5d0da549c7f3" (UID: "95cd3d5a-1000-4212-b4d9-5d0da549c7f3"). InnerVolumeSpecName "pvc-f33acc08-4fd4-473f-9263-ec4841f48809". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.338204 4771 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-f33acc08-4fd4-473f-9263-ec4841f48809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f33acc08-4fd4-473f-9263-ec4841f48809\") on node \"crc\" " Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.384312 4771 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.384558 4771 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-f33acc08-4fd4-473f-9263-ec4841f48809" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f33acc08-4fd4-473f-9263-ec4841f48809") on node "crc" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.447910 4771 reconciler_common.go:293] "Volume detached for volume \"pvc-f33acc08-4fd4-473f-9263-ec4841f48809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f33acc08-4fd4-473f-9263-ec4841f48809\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.455352 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"95cd3d5a-1000-4212-b4d9-5d0da549c7f3","Type":"ContainerDied","Data":"5760d3a33afe8aa8fab457b68102811193f6ca9901b26714b020c370de834fa6"} Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.455424 4771 scope.go:117] "RemoveContainer" containerID="61c3b16f18cbb70283e079c3cf0a250091cf91ceb3e801c047cf6918fb937be1" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.455614 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.520659 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.540891 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.559838 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 09:59:30 crc kubenswrapper[4771]: E1002 09:59:30.560540 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="init-config-reloader" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.560562 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="init-config-reloader" Oct 02 09:59:30 crc kubenswrapper[4771]: E1002 09:59:30.560591 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="thanos-sidecar" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.560598 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="thanos-sidecar" Oct 02 09:59:30 crc kubenswrapper[4771]: E1002 09:59:30.560611 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="prometheus" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.560617 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="prometheus" Oct 02 09:59:30 crc kubenswrapper[4771]: E1002 09:59:30.560627 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="config-reloader" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.560633 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="config-reloader" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.560848 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="config-reloader" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.560881 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="thanos-sidecar" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.560894 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" containerName="prometheus" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.563094 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.568557 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.568557 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.568720 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.568796 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.568872 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-656s8" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.568932 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.573913 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.576528 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.657013 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.659600 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.660076 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.660216 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.660324 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-config\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.660668 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6caafe76-3bdf-423b-abb6-c2057f835d38-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.660754 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6caafe76-3bdf-423b-abb6-c2057f835d38-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.660950 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f33acc08-4fd4-473f-9263-ec4841f48809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f33acc08-4fd4-473f-9263-ec4841f48809\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.660998 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5kkz\" (UniqueName: \"kubernetes.io/projected/6caafe76-3bdf-423b-abb6-c2057f835d38-kube-api-access-b5kkz\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.661062 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.661216 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6caafe76-3bdf-423b-abb6-c2057f835d38-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.762961 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f33acc08-4fd4-473f-9263-ec4841f48809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f33acc08-4fd4-473f-9263-ec4841f48809\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.763050 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5kkz\" (UniqueName: \"kubernetes.io/projected/6caafe76-3bdf-423b-abb6-c2057f835d38-kube-api-access-b5kkz\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.763092 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.763209 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6caafe76-3bdf-423b-abb6-c2057f835d38-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.763285 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.763327 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.763386 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.763415 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.763444 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-config\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.763487 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6caafe76-3bdf-423b-abb6-c2057f835d38-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.763528 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6caafe76-3bdf-423b-abb6-c2057f835d38-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.765103 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6caafe76-3bdf-423b-abb6-c2057f835d38-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.770171 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6caafe76-3bdf-423b-abb6-c2057f835d38-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.770184 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.770735 4771 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.770770 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.770790 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f33acc08-4fd4-473f-9263-ec4841f48809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f33acc08-4fd4-473f-9263-ec4841f48809\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5158297136c2a151404177ed2b564800e7d4895241ff310a5747c375183a3a49/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.771027 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.771303 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-config\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.772547 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6caafe76-3bdf-423b-abb6-c2057f835d38-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.773169 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.776895 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6caafe76-3bdf-423b-abb6-c2057f835d38-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.784103 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5kkz\" (UniqueName: \"kubernetes.io/projected/6caafe76-3bdf-423b-abb6-c2057f835d38-kube-api-access-b5kkz\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.822319 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f33acc08-4fd4-473f-9263-ec4841f48809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f33acc08-4fd4-473f-9263-ec4841f48809\") pod \"prometheus-metric-storage-0\" (UID: \"6caafe76-3bdf-423b-abb6-c2057f835d38\") " pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:30 crc kubenswrapper[4771]: I1002 09:59:30.890291 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 09:59:31 crc kubenswrapper[4771]: I1002 09:59:31.697673 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95cd3d5a-1000-4212-b4d9-5d0da549c7f3" path="/var/lib/kubelet/pods/95cd3d5a-1000-4212-b4d9-5d0da549c7f3/volumes" Oct 02 09:59:42 crc kubenswrapper[4771]: I1002 09:59:42.145996 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:59:42 crc kubenswrapper[4771]: I1002 09:59:42.146673 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:59:42 crc kubenswrapper[4771]: I1002 09:59:42.146733 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 09:59:42 crc kubenswrapper[4771]: I1002 09:59:42.147625 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"405cf9bc15cbb64122966859b742b139fd0128a50163288165936c501ca2e3e6"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:59:42 crc kubenswrapper[4771]: I1002 09:59:42.147683 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://405cf9bc15cbb64122966859b742b139fd0128a50163288165936c501ca2e3e6" gracePeriod=600 Oct 02 09:59:44 crc kubenswrapper[4771]: I1002 09:59:44.634421 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="405cf9bc15cbb64122966859b742b139fd0128a50163288165936c501ca2e3e6" exitCode=0 Oct 02 09:59:44 crc kubenswrapper[4771]: I1002 09:59:44.634626 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"405cf9bc15cbb64122966859b742b139fd0128a50163288165936c501ca2e3e6"} Oct 02 09:59:45 crc kubenswrapper[4771]: I1002 09:59:45.350988 4771 scope.go:117] "RemoveContainer" containerID="7e88704fb0629cfd87c9422d98c45c8ecb8684b1aa2ddf1d73efee5edce3e523" Oct 02 09:59:45 crc kubenswrapper[4771]: I1002 09:59:45.540960 4771 scope.go:117] "RemoveContainer" containerID="c3286830dd0529949a1d53aeff88ff2c8e1154091dff38d555badaf1e6af5d75" Oct 02 09:59:45 crc kubenswrapper[4771]: I1002 09:59:45.623587 4771 scope.go:117] "RemoveContainer" containerID="6fd70e48c855a5f9bf81cbf37919a8679ecbde45a91dccd0e2a6fc1a67916fa6" Oct 02 09:59:45 crc kubenswrapper[4771]: I1002 09:59:45.744182 4771 scope.go:117] "RemoveContainer" containerID="9385c2320866e6d26f67de876609b5bcc62a83b76b282bedf5fb7cfa0fec4a42" Oct 02 09:59:45 crc kubenswrapper[4771]: I1002 09:59:45.801575 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-73f2-account-create-phct2"] Oct 02 09:59:45 crc kubenswrapper[4771]: W1002 09:59:45.825766 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb63a764e_b703_434d_860d_7774c43035aa.slice/crio-de24350d24953de96440b7efe33a06de5a2b594a8b9d3ee2b60ae705352ccaba WatchSource:0}: Error finding container de24350d24953de96440b7efe33a06de5a2b594a8b9d3ee2b60ae705352ccaba: Status 404 returned error can't find the container with id de24350d24953de96440b7efe33a06de5a2b594a8b9d3ee2b60ae705352ccaba Oct 02 09:59:46 crc kubenswrapper[4771]: I1002 09:59:46.262607 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-4b06-account-create-5s4cn"] Oct 02 09:59:46 crc kubenswrapper[4771]: I1002 09:59:46.273654 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9938-account-create-jj42p"] Oct 02 09:59:46 crc kubenswrapper[4771]: I1002 09:59:46.289975 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-816a-account-create-6pk4g"] Oct 02 09:59:46 crc kubenswrapper[4771]: I1002 09:59:46.342309 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 09:59:46 crc kubenswrapper[4771]: I1002 09:59:46.675039 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-816a-account-create-6pk4g" event={"ID":"6cc9f109-903c-4e52-8031-56898f9128ad","Type":"ContainerStarted","Data":"b43331380a51e9e108523817f79e5cee8b62bb341db703dd4d8a5b393a1d29e0"} Oct 02 09:59:46 crc kubenswrapper[4771]: I1002 09:59:46.678096 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"2ed16dff4f8c412a4724b6bc7ff859ddf10fd7debcade0aabe130d60c75a3058"} Oct 02 09:59:46 crc kubenswrapper[4771]: I1002 09:59:46.684550 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9938-account-create-jj42p" event={"ID":"e215bc59-d489-4b95-8348-fe23e36f2d6e","Type":"ContainerStarted","Data":"5f82164bc7fef559562ce66a0378c597156a7971c93cc90be5a7af178d98a57d"} Oct 02 09:59:46 crc kubenswrapper[4771]: I1002 09:59:46.686542 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6caafe76-3bdf-423b-abb6-c2057f835d38","Type":"ContainerStarted","Data":"b0a93335c0c56e2208059288d4e8099fc96f25d825c657d2a4f746e259d32d7c"} Oct 02 09:59:46 crc kubenswrapper[4771]: I1002 09:59:46.687952 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-4b06-account-create-5s4cn" event={"ID":"fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d","Type":"ContainerStarted","Data":"bae8f36b3f0759bb4c3aa004d20a91549e9ef4df129137311d93bd57871f3e05"} Oct 02 09:59:46 crc kubenswrapper[4771]: I1002 09:59:46.689632 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-73f2-account-create-phct2" event={"ID":"b63a764e-b703-434d-860d-7774c43035aa","Type":"ContainerStarted","Data":"b0039d8d6fa5b65a9b4acc8877a87b725d441cab8d55a36a7c6b8e7516123710"} Oct 02 09:59:46 crc kubenswrapper[4771]: I1002 09:59:46.689773 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-73f2-account-create-phct2" event={"ID":"b63a764e-b703-434d-860d-7774c43035aa","Type":"ContainerStarted","Data":"de24350d24953de96440b7efe33a06de5a2b594a8b9d3ee2b60ae705352ccaba"} Oct 02 09:59:47 crc kubenswrapper[4771]: I1002 09:59:47.721926 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-816a-account-create-6pk4g" event={"ID":"6cc9f109-903c-4e52-8031-56898f9128ad","Type":"ContainerStarted","Data":"d963749382478e971ca61a3a6dc2f234cad4fabff4235f72d1f4cc6c8d0890be"} Oct 02 09:59:47 crc kubenswrapper[4771]: I1002 09:59:47.731828 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9938-account-create-jj42p" event={"ID":"e215bc59-d489-4b95-8348-fe23e36f2d6e","Type":"ContainerStarted","Data":"a5676c418f62a7deab378dac49f7fcd0cb147b8c3f19426fdac362056bbea66e"} Oct 02 09:59:47 crc kubenswrapper[4771]: I1002 09:59:47.735679 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-4b06-account-create-5s4cn" event={"ID":"fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d","Type":"ContainerStarted","Data":"079dd3d001178298098973e0fad451f6df3e8f546789da08a53b73571170ae9f"} Oct 02 09:59:47 crc kubenswrapper[4771]: I1002 09:59:47.740459 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772"} Oct 02 09:59:47 crc kubenswrapper[4771]: I1002 09:59:47.762072 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-73f2-account-create-phct2" podStartSLOduration=21.762016329 podStartE2EDuration="21.762016329s" podCreationTimestamp="2025-10-02 09:59:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:59:47.758113159 +0000 UTC m=+1375.405798226" watchObservedRunningTime="2025-10-02 09:59:47.762016329 +0000 UTC m=+1375.409701396" Oct 02 09:59:48 crc kubenswrapper[4771]: I1002 09:59:48.761247 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"58aaf808421d916fd77f8e5c61b16d36d0571d5f8052ca82f08b9437392f0e1e"} Oct 02 09:59:48 crc kubenswrapper[4771]: I1002 09:59:48.791836 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-4b06-account-create-5s4cn" podStartSLOduration=22.791813051 podStartE2EDuration="22.791813051s" podCreationTimestamp="2025-10-02 09:59:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:59:48.77923666 +0000 UTC m=+1376.426921717" watchObservedRunningTime="2025-10-02 09:59:48.791813051 +0000 UTC m=+1376.439498118" Oct 02 09:59:48 crc kubenswrapper[4771]: I1002 09:59:48.806612 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-816a-account-create-6pk4g" podStartSLOduration=22.806576769 podStartE2EDuration="22.806576769s" podCreationTimestamp="2025-10-02 09:59:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:59:48.800921544 +0000 UTC m=+1376.448606611" watchObservedRunningTime="2025-10-02 09:59:48.806576769 +0000 UTC m=+1376.454261836" Oct 02 09:59:48 crc kubenswrapper[4771]: I1002 09:59:48.854845 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-9938-account-create-jj42p" podStartSLOduration=22.854820612 podStartE2EDuration="22.854820612s" podCreationTimestamp="2025-10-02 09:59:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:59:48.850057881 +0000 UTC m=+1376.497742948" watchObservedRunningTime="2025-10-02 09:59:48.854820612 +0000 UTC m=+1376.502505679" Oct 02 09:59:48 crc kubenswrapper[4771]: E1002 09:59:48.999507 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-keystone:current-podified" Oct 02 09:59:48 crc kubenswrapper[4771]: E1002 09:59:48.999744 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:keystone-db-sync,Image:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,Command:[/bin/bash],Args:[-c keystone-manage db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/keystone/keystone.conf,SubPath:keystone.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mbfbp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42425,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42425,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-db-sync-cjqrx_openstack(9f0f7ed9-b19d-41e3-9eb3-d6431979e26a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:59:49 crc kubenswrapper[4771]: E1002 09:59:49.000989 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/keystone-db-sync-cjqrx" podUID="9f0f7ed9-b19d-41e3-9eb3-d6431979e26a" Oct 02 09:59:49 crc kubenswrapper[4771]: I1002 09:59:49.776876 4771 generic.go:334] "Generic (PLEG): container finished" podID="6cc9f109-903c-4e52-8031-56898f9128ad" containerID="d963749382478e971ca61a3a6dc2f234cad4fabff4235f72d1f4cc6c8d0890be" exitCode=0 Oct 02 09:59:49 crc kubenswrapper[4771]: I1002 09:59:49.777320 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-816a-account-create-6pk4g" event={"ID":"6cc9f109-903c-4e52-8031-56898f9128ad","Type":"ContainerDied","Data":"d963749382478e971ca61a3a6dc2f234cad4fabff4235f72d1f4cc6c8d0890be"} Oct 02 09:59:49 crc kubenswrapper[4771]: I1002 09:59:49.780245 4771 generic.go:334] "Generic (PLEG): container finished" podID="e215bc59-d489-4b95-8348-fe23e36f2d6e" containerID="a5676c418f62a7deab378dac49f7fcd0cb147b8c3f19426fdac362056bbea66e" exitCode=0 Oct 02 09:59:49 crc kubenswrapper[4771]: I1002 09:59:49.780292 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9938-account-create-jj42p" event={"ID":"e215bc59-d489-4b95-8348-fe23e36f2d6e","Type":"ContainerDied","Data":"a5676c418f62a7deab378dac49f7fcd0cb147b8c3f19426fdac362056bbea66e"} Oct 02 09:59:49 crc kubenswrapper[4771]: I1002 09:59:49.782646 4771 generic.go:334] "Generic (PLEG): container finished" podID="fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d" containerID="079dd3d001178298098973e0fad451f6df3e8f546789da08a53b73571170ae9f" exitCode=0 Oct 02 09:59:49 crc kubenswrapper[4771]: I1002 09:59:49.782716 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-4b06-account-create-5s4cn" event={"ID":"fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d","Type":"ContainerDied","Data":"079dd3d001178298098973e0fad451f6df3e8f546789da08a53b73571170ae9f"} Oct 02 09:59:49 crc kubenswrapper[4771]: I1002 09:59:49.787557 4771 generic.go:334] "Generic (PLEG): container finished" podID="b63a764e-b703-434d-860d-7774c43035aa" containerID="b0039d8d6fa5b65a9b4acc8877a87b725d441cab8d55a36a7c6b8e7516123710" exitCode=0 Oct 02 09:59:49 crc kubenswrapper[4771]: I1002 09:59:49.788385 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-73f2-account-create-phct2" event={"ID":"b63a764e-b703-434d-860d-7774c43035aa","Type":"ContainerDied","Data":"b0039d8d6fa5b65a9b4acc8877a87b725d441cab8d55a36a7c6b8e7516123710"} Oct 02 09:59:49 crc kubenswrapper[4771]: E1002 09:59:49.791230 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-keystone:current-podified\\\"\"" pod="openstack/keystone-db-sync-cjqrx" podUID="9f0f7ed9-b19d-41e3-9eb3-d6431979e26a" Oct 02 09:59:50 crc kubenswrapper[4771]: I1002 09:59:50.848291 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"e3f1498d239345bee59819486fbd8b79b1c7356cf967a2710c89687056cf7e61"} Oct 02 09:59:50 crc kubenswrapper[4771]: I1002 09:59:50.849040 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"570f1bbf76d930b0c8ff51dc4a07267de60d77f7fee6c813a1ebc38c5d6d0463"} Oct 02 09:59:50 crc kubenswrapper[4771]: I1002 09:59:50.866728 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6caafe76-3bdf-423b-abb6-c2057f835d38","Type":"ContainerStarted","Data":"bc85c78471fb43c2f5f22c35222cc1b6fac30f84d2c03fa572a54e3f8fde5e6f"} Oct 02 09:59:51 crc kubenswrapper[4771]: I1002 09:59:51.775747 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9938-account-create-jj42p" Oct 02 09:59:51 crc kubenswrapper[4771]: I1002 09:59:51.886556 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"879c83886ba807f6e0eea0363e60fb606baba1d922b7d092e6111a7cd18c2856"} Oct 02 09:59:51 crc kubenswrapper[4771]: I1002 09:59:51.886620 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"51ea1670ccde3365a50f3a50576d52c989e101085589e2c84236a4734bcd1581"} Oct 02 09:59:51 crc kubenswrapper[4771]: I1002 09:59:51.898845 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9938-account-create-jj42p" Oct 02 09:59:51 crc kubenswrapper[4771]: I1002 09:59:51.898857 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9938-account-create-jj42p" event={"ID":"e215bc59-d489-4b95-8348-fe23e36f2d6e","Type":"ContainerDied","Data":"5f82164bc7fef559562ce66a0378c597156a7971c93cc90be5a7af178d98a57d"} Oct 02 09:59:51 crc kubenswrapper[4771]: I1002 09:59:51.898932 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f82164bc7fef559562ce66a0378c597156a7971c93cc90be5a7af178d98a57d" Oct 02 09:59:51 crc kubenswrapper[4771]: I1002 09:59:51.958407 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2ms\" (UniqueName: \"kubernetes.io/projected/e215bc59-d489-4b95-8348-fe23e36f2d6e-kube-api-access-bf2ms\") pod \"e215bc59-d489-4b95-8348-fe23e36f2d6e\" (UID: \"e215bc59-d489-4b95-8348-fe23e36f2d6e\") " Oct 02 09:59:51 crc kubenswrapper[4771]: I1002 09:59:51.967271 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e215bc59-d489-4b95-8348-fe23e36f2d6e-kube-api-access-bf2ms" (OuterVolumeSpecName: "kube-api-access-bf2ms") pod "e215bc59-d489-4b95-8348-fe23e36f2d6e" (UID: "e215bc59-d489-4b95-8348-fe23e36f2d6e"). InnerVolumeSpecName "kube-api-access-bf2ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.061950 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2ms\" (UniqueName: \"kubernetes.io/projected/e215bc59-d489-4b95-8348-fe23e36f2d6e-kube-api-access-bf2ms\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.084012 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-816a-account-create-6pk4g" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.095713 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-73f2-account-create-phct2" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.111783 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-4b06-account-create-5s4cn" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.163778 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdc94\" (UniqueName: \"kubernetes.io/projected/6cc9f109-903c-4e52-8031-56898f9128ad-kube-api-access-qdc94\") pod \"6cc9f109-903c-4e52-8031-56898f9128ad\" (UID: \"6cc9f109-903c-4e52-8031-56898f9128ad\") " Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.171930 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cc9f109-903c-4e52-8031-56898f9128ad-kube-api-access-qdc94" (OuterVolumeSpecName: "kube-api-access-qdc94") pod "6cc9f109-903c-4e52-8031-56898f9128ad" (UID: "6cc9f109-903c-4e52-8031-56898f9128ad"). InnerVolumeSpecName "kube-api-access-qdc94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.267993 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4mhb\" (UniqueName: \"kubernetes.io/projected/fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d-kube-api-access-m4mhb\") pod \"fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d\" (UID: \"fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d\") " Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.268095 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh8fb\" (UniqueName: \"kubernetes.io/projected/b63a764e-b703-434d-860d-7774c43035aa-kube-api-access-xh8fb\") pod \"b63a764e-b703-434d-860d-7774c43035aa\" (UID: \"b63a764e-b703-434d-860d-7774c43035aa\") " Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.268883 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdc94\" (UniqueName: \"kubernetes.io/projected/6cc9f109-903c-4e52-8031-56898f9128ad-kube-api-access-qdc94\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.273529 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d-kube-api-access-m4mhb" (OuterVolumeSpecName: "kube-api-access-m4mhb") pod "fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d" (UID: "fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d"). InnerVolumeSpecName "kube-api-access-m4mhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.275070 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b63a764e-b703-434d-860d-7774c43035aa-kube-api-access-xh8fb" (OuterVolumeSpecName: "kube-api-access-xh8fb") pod "b63a764e-b703-434d-860d-7774c43035aa" (UID: "b63a764e-b703-434d-860d-7774c43035aa"). InnerVolumeSpecName "kube-api-access-xh8fb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.372478 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4mhb\" (UniqueName: \"kubernetes.io/projected/fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d-kube-api-access-m4mhb\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.372537 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh8fb\" (UniqueName: \"kubernetes.io/projected/b63a764e-b703-434d-860d-7774c43035aa-kube-api-access-xh8fb\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.911494 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-816a-account-create-6pk4g" event={"ID":"6cc9f109-903c-4e52-8031-56898f9128ad","Type":"ContainerDied","Data":"b43331380a51e9e108523817f79e5cee8b62bb341db703dd4d8a5b393a1d29e0"} Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.912090 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b43331380a51e9e108523817f79e5cee8b62bb341db703dd4d8a5b393a1d29e0" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.912257 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-816a-account-create-6pk4g" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.923469 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"251a5773581a5543759815dadbb886ce8e28ad25db408dbb16c8e5c73c2ca34e"} Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.929149 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-4b06-account-create-5s4cn" event={"ID":"fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d","Type":"ContainerDied","Data":"bae8f36b3f0759bb4c3aa004d20a91549e9ef4df129137311d93bd57871f3e05"} Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.929313 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bae8f36b3f0759bb4c3aa004d20a91549e9ef4df129137311d93bd57871f3e05" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.929226 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-4b06-account-create-5s4cn" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.931421 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-73f2-account-create-phct2" event={"ID":"b63a764e-b703-434d-860d-7774c43035aa","Type":"ContainerDied","Data":"de24350d24953de96440b7efe33a06de5a2b594a8b9d3ee2b60ae705352ccaba"} Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.931478 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de24350d24953de96440b7efe33a06de5a2b594a8b9d3ee2b60ae705352ccaba" Oct 02 09:59:52 crc kubenswrapper[4771]: I1002 09:59:52.931583 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-73f2-account-create-phct2" Oct 02 09:59:53 crc kubenswrapper[4771]: I1002 09:59:53.947108 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"37c1ca014ca238c9c7ba9e1e65eae4ff1d94f3cc06d7c6be4d7eaa902397dba5"} Oct 02 09:59:53 crc kubenswrapper[4771]: I1002 09:59:53.947561 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"b78d8b4aaa7f61edd99007f681a110c6b0426c8de789234f473e9bd5448a8b0a"} Oct 02 09:59:53 crc kubenswrapper[4771]: I1002 09:59:53.947572 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"79b32c151d93cccaeccdba8a59131d048946814865a7c9379653be10322e29d7"} Oct 02 09:59:53 crc kubenswrapper[4771]: I1002 09:59:53.947583 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"c56a91065ffedc7d303172cbc4f4c7115cd5c7a358ac2a798e7a75162de536d1"} Oct 02 09:59:54 crc kubenswrapper[4771]: I1002 09:59:54.968566 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"9a3fda3564141b32296e77b877b82b9f92ad35161bfbe8da10616af0c318aadb"} Oct 02 09:59:54 crc kubenswrapper[4771]: I1002 09:59:54.968985 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4c4deacb-7b2f-49d1-8c7e-b87913f027d1","Type":"ContainerStarted","Data":"c5af386fcd9b081933c7352d9ee2308cc5dfbe3178eb8a8ba59e6f1f55bb3653"} Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.013680 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=41.128085695 podStartE2EDuration="1m17.013655239s" podCreationTimestamp="2025-10-02 09:58:38 +0000 UTC" firstStartedPulling="2025-10-02 09:59:16.68098944 +0000 UTC m=+1344.328674507" lastFinishedPulling="2025-10-02 09:59:52.566558984 +0000 UTC m=+1380.214244051" observedRunningTime="2025-10-02 09:59:55.004028813 +0000 UTC m=+1382.651713880" watchObservedRunningTime="2025-10-02 09:59:55.013655239 +0000 UTC m=+1382.661340316" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.303582 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-kjwbj"] Oct 02 09:59:55 crc kubenswrapper[4771]: E1002 09:59:55.304296 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d" containerName="mariadb-account-create" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.304322 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d" containerName="mariadb-account-create" Oct 02 09:59:55 crc kubenswrapper[4771]: E1002 09:59:55.304336 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cc9f109-903c-4e52-8031-56898f9128ad" containerName="mariadb-account-create" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.304343 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cc9f109-903c-4e52-8031-56898f9128ad" containerName="mariadb-account-create" Oct 02 09:59:55 crc kubenswrapper[4771]: E1002 09:59:55.304381 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b63a764e-b703-434d-860d-7774c43035aa" containerName="mariadb-account-create" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.304389 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b63a764e-b703-434d-860d-7774c43035aa" containerName="mariadb-account-create" Oct 02 09:59:55 crc kubenswrapper[4771]: E1002 09:59:55.304410 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e215bc59-d489-4b95-8348-fe23e36f2d6e" containerName="mariadb-account-create" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.304418 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e215bc59-d489-4b95-8348-fe23e36f2d6e" containerName="mariadb-account-create" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.304644 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="e215bc59-d489-4b95-8348-fe23e36f2d6e" containerName="mariadb-account-create" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.304668 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d" containerName="mariadb-account-create" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.304685 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cc9f109-903c-4e52-8031-56898f9128ad" containerName="mariadb-account-create" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.304705 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="b63a764e-b703-434d-860d-7774c43035aa" containerName="mariadb-account-create" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.306325 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.322551 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.323495 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-kjwbj"] Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.472832 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-config\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.472934 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwjgf\" (UniqueName: \"kubernetes.io/projected/d9d66796-ff80-4c4b-a753-f13cb10e2754-kube-api-access-gwjgf\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.472980 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.473230 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.473620 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-dns-svc\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.473685 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.575449 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.575530 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.575660 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-dns-svc\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.575693 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.575810 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-config\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.575858 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwjgf\" (UniqueName: \"kubernetes.io/projected/d9d66796-ff80-4c4b-a753-f13cb10e2754-kube-api-access-gwjgf\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.578435 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.578818 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.578926 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-dns-svc\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.579297 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-config\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.579776 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.600540 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwjgf\" (UniqueName: \"kubernetes.io/projected/d9d66796-ff80-4c4b-a753-f13cb10e2754-kube-api-access-gwjgf\") pod \"dnsmasq-dns-764c5664d7-kjwbj\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:55 crc kubenswrapper[4771]: I1002 09:59:55.653493 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:56 crc kubenswrapper[4771]: I1002 09:59:56.240231 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-kjwbj"] Oct 02 09:59:57 crc kubenswrapper[4771]: I1002 09:59:57.003289 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" event={"ID":"d9d66796-ff80-4c4b-a753-f13cb10e2754","Type":"ContainerStarted","Data":"489d8258995a5b34243e22f381ff037037aed40983bda64927b307c8efc16d0a"} Oct 02 09:59:57 crc kubenswrapper[4771]: I1002 09:59:57.003752 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" event={"ID":"d9d66796-ff80-4c4b-a753-f13cb10e2754","Type":"ContainerStarted","Data":"8b3f9bbe85b0a45f6659e13d7f15c56cd09079547f2716e6c34957fa689e26cc"} Oct 02 09:59:58 crc kubenswrapper[4771]: I1002 09:59:58.018400 4771 generic.go:334] "Generic (PLEG): container finished" podID="d9d66796-ff80-4c4b-a753-f13cb10e2754" containerID="489d8258995a5b34243e22f381ff037037aed40983bda64927b307c8efc16d0a" exitCode=0 Oct 02 09:59:58 crc kubenswrapper[4771]: I1002 09:59:58.018486 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" event={"ID":"d9d66796-ff80-4c4b-a753-f13cb10e2754","Type":"ContainerDied","Data":"489d8258995a5b34243e22f381ff037037aed40983bda64927b307c8efc16d0a"} Oct 02 09:59:59 crc kubenswrapper[4771]: I1002 09:59:59.033379 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" event={"ID":"d9d66796-ff80-4c4b-a753-f13cb10e2754","Type":"ContainerStarted","Data":"db31a9ebdb450dfb16c11ed4a92b24c31374708125c0aece21d9333d8d67f7cb"} Oct 02 09:59:59 crc kubenswrapper[4771]: I1002 09:59:59.033923 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 09:59:59 crc kubenswrapper[4771]: I1002 09:59:59.061020 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" podStartSLOduration=4.060993393 podStartE2EDuration="4.060993393s" podCreationTimestamp="2025-10-02 09:59:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:59:59.053424159 +0000 UTC m=+1386.701109256" watchObservedRunningTime="2025-10-02 09:59:59.060993393 +0000 UTC m=+1386.708678460" Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.050170 4771 generic.go:334] "Generic (PLEG): container finished" podID="4c6203ef-b851-49ba-ae59-9b62c473c497" containerID="d66e976c25ca5a30398fd5404e8f0a51209b789267b8e1d2ab18533250acb02f" exitCode=0 Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.056598 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-t42dx" event={"ID":"4c6203ef-b851-49ba-ae59-9b62c473c497","Type":"ContainerDied","Data":"d66e976c25ca5a30398fd5404e8f0a51209b789267b8e1d2ab18533250acb02f"} Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.143247 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7"] Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.146336 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.149421 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.149960 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.163443 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7"] Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.299295 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/325b9e47-f103-41d4-ad1f-f9fb2eecde45-config-volume\") pod \"collect-profiles-29323320-gqsw7\" (UID: \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.299353 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/325b9e47-f103-41d4-ad1f-f9fb2eecde45-secret-volume\") pod \"collect-profiles-29323320-gqsw7\" (UID: \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.299404 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vj26\" (UniqueName: \"kubernetes.io/projected/325b9e47-f103-41d4-ad1f-f9fb2eecde45-kube-api-access-9vj26\") pod \"collect-profiles-29323320-gqsw7\" (UID: \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.401084 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/325b9e47-f103-41d4-ad1f-f9fb2eecde45-config-volume\") pod \"collect-profiles-29323320-gqsw7\" (UID: \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.402247 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/325b9e47-f103-41d4-ad1f-f9fb2eecde45-secret-volume\") pod \"collect-profiles-29323320-gqsw7\" (UID: \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.402356 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vj26\" (UniqueName: \"kubernetes.io/projected/325b9e47-f103-41d4-ad1f-f9fb2eecde45-kube-api-access-9vj26\") pod \"collect-profiles-29323320-gqsw7\" (UID: \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.402205 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/325b9e47-f103-41d4-ad1f-f9fb2eecde45-config-volume\") pod \"collect-profiles-29323320-gqsw7\" (UID: \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.413687 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/325b9e47-f103-41d4-ad1f-f9fb2eecde45-secret-volume\") pod \"collect-profiles-29323320-gqsw7\" (UID: \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.438597 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vj26\" (UniqueName: \"kubernetes.io/projected/325b9e47-f103-41d4-ad1f-f9fb2eecde45-kube-api-access-9vj26\") pod \"collect-profiles-29323320-gqsw7\" (UID: \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" Oct 02 10:00:00 crc kubenswrapper[4771]: I1002 10:00:00.474554 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.002978 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7"] Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.087754 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" event={"ID":"325b9e47-f103-41d4-ad1f-f9fb2eecde45","Type":"ContainerStarted","Data":"5f081a1501b0b695ef271e7492ff39c8712c434a3f9b07f930adb6aa05d0ff72"} Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.749170 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-t42dx" Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.838411 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-db-sync-config-data\") pod \"4c6203ef-b851-49ba-ae59-9b62c473c497\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.838485 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-combined-ca-bundle\") pod \"4c6203ef-b851-49ba-ae59-9b62c473c497\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.838708 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-config-data\") pod \"4c6203ef-b851-49ba-ae59-9b62c473c497\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.838843 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rx4jk\" (UniqueName: \"kubernetes.io/projected/4c6203ef-b851-49ba-ae59-9b62c473c497-kube-api-access-rx4jk\") pod \"4c6203ef-b851-49ba-ae59-9b62c473c497\" (UID: \"4c6203ef-b851-49ba-ae59-9b62c473c497\") " Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.847245 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4c6203ef-b851-49ba-ae59-9b62c473c497" (UID: "4c6203ef-b851-49ba-ae59-9b62c473c497"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.851758 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c6203ef-b851-49ba-ae59-9b62c473c497-kube-api-access-rx4jk" (OuterVolumeSpecName: "kube-api-access-rx4jk") pod "4c6203ef-b851-49ba-ae59-9b62c473c497" (UID: "4c6203ef-b851-49ba-ae59-9b62c473c497"). InnerVolumeSpecName "kube-api-access-rx4jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.899670 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c6203ef-b851-49ba-ae59-9b62c473c497" (UID: "4c6203ef-b851-49ba-ae59-9b62c473c497"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.915537 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-config-data" (OuterVolumeSpecName: "config-data") pod "4c6203ef-b851-49ba-ae59-9b62c473c497" (UID: "4c6203ef-b851-49ba-ae59-9b62c473c497"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.941171 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.941222 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rx4jk\" (UniqueName: \"kubernetes.io/projected/4c6203ef-b851-49ba-ae59-9b62c473c497-kube-api-access-rx4jk\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.941238 4771 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:01 crc kubenswrapper[4771]: I1002 10:00:01.941248 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c6203ef-b851-49ba-ae59-9b62c473c497-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.131995 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-t42dx" event={"ID":"4c6203ef-b851-49ba-ae59-9b62c473c497","Type":"ContainerDied","Data":"1a3e64e0b57fc0ff5ac8675f92580dd7ac4dc3190e3e759bb322291a080db9fc"} Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.132072 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a3e64e0b57fc0ff5ac8675f92580dd7ac4dc3190e3e759bb322291a080db9fc" Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.132237 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-t42dx" Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.143111 4771 generic.go:334] "Generic (PLEG): container finished" podID="325b9e47-f103-41d4-ad1f-f9fb2eecde45" containerID="e138a2ca1b3d6a52d4b4f986aa194b1a52fe0f83d5004be678043c565d1242dd" exitCode=0 Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.143202 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" event={"ID":"325b9e47-f103-41d4-ad1f-f9fb2eecde45","Type":"ContainerDied","Data":"e138a2ca1b3d6a52d4b4f986aa194b1a52fe0f83d5004be678043c565d1242dd"} Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.156217 4771 generic.go:334] "Generic (PLEG): container finished" podID="6caafe76-3bdf-423b-abb6-c2057f835d38" containerID="bc85c78471fb43c2f5f22c35222cc1b6fac30f84d2c03fa572a54e3f8fde5e6f" exitCode=0 Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.156325 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6caafe76-3bdf-423b-abb6-c2057f835d38","Type":"ContainerDied","Data":"bc85c78471fb43c2f5f22c35222cc1b6fac30f84d2c03fa572a54e3f8fde5e6f"} Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.711511 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-kjwbj"] Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.712324 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" podUID="d9d66796-ff80-4c4b-a753-f13cb10e2754" containerName="dnsmasq-dns" containerID="cri-o://db31a9ebdb450dfb16c11ed4a92b24c31374708125c0aece21d9333d8d67f7cb" gracePeriod=10 Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.785850 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-x9s6s"] Oct 02 10:00:02 crc kubenswrapper[4771]: E1002 10:00:02.786656 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c6203ef-b851-49ba-ae59-9b62c473c497" containerName="glance-db-sync" Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.786685 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c6203ef-b851-49ba-ae59-9b62c473c497" containerName="glance-db-sync" Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.786947 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c6203ef-b851-49ba-ae59-9b62c473c497" containerName="glance-db-sync" Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.788601 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.799114 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-x9s6s"] Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.933150 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdmk7\" (UniqueName: \"kubernetes.io/projected/5dcb591a-9c0b-44ce-a0f0-db26883d5479-kube-api-access-kdmk7\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.934589 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-config\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.934966 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.935097 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.935747 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:02 crc kubenswrapper[4771]: I1002 10:00:02.935874 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.043823 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdmk7\" (UniqueName: \"kubernetes.io/projected/5dcb591a-9c0b-44ce-a0f0-db26883d5479-kube-api-access-kdmk7\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.043888 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-config\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.043993 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.044041 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.044108 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.044156 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.052482 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.053085 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.053472 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-config\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.054114 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.054747 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.070029 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdmk7\" (UniqueName: \"kubernetes.io/projected/5dcb591a-9c0b-44ce-a0f0-db26883d5479-kube-api-access-kdmk7\") pod \"dnsmasq-dns-74f6bcbc87-x9s6s\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.189092 4771 generic.go:334] "Generic (PLEG): container finished" podID="d9d66796-ff80-4c4b-a753-f13cb10e2754" containerID="db31a9ebdb450dfb16c11ed4a92b24c31374708125c0aece21d9333d8d67f7cb" exitCode=0 Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.189265 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" event={"ID":"d9d66796-ff80-4c4b-a753-f13cb10e2754","Type":"ContainerDied","Data":"db31a9ebdb450dfb16c11ed4a92b24c31374708125c0aece21d9333d8d67f7cb"} Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.208172 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6caafe76-3bdf-423b-abb6-c2057f835d38","Type":"ContainerStarted","Data":"a7c785568f133b9a9c6fcaab4da524b66a37506cbffa890c7271fb558ddbc814"} Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.226774 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.350774 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.459168 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwjgf\" (UniqueName: \"kubernetes.io/projected/d9d66796-ff80-4c4b-a753-f13cb10e2754-kube-api-access-gwjgf\") pod \"d9d66796-ff80-4c4b-a753-f13cb10e2754\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.459306 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-config\") pod \"d9d66796-ff80-4c4b-a753-f13cb10e2754\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.459403 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-dns-svc\") pod \"d9d66796-ff80-4c4b-a753-f13cb10e2754\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.459505 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-ovsdbserver-nb\") pod \"d9d66796-ff80-4c4b-a753-f13cb10e2754\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.459537 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-dns-swift-storage-0\") pod \"d9d66796-ff80-4c4b-a753-f13cb10e2754\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.459597 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-ovsdbserver-sb\") pod \"d9d66796-ff80-4c4b-a753-f13cb10e2754\" (UID: \"d9d66796-ff80-4c4b-a753-f13cb10e2754\") " Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.483220 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9d66796-ff80-4c4b-a753-f13cb10e2754-kube-api-access-gwjgf" (OuterVolumeSpecName: "kube-api-access-gwjgf") pod "d9d66796-ff80-4c4b-a753-f13cb10e2754" (UID: "d9d66796-ff80-4c4b-a753-f13cb10e2754"). InnerVolumeSpecName "kube-api-access-gwjgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.549801 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d9d66796-ff80-4c4b-a753-f13cb10e2754" (UID: "d9d66796-ff80-4c4b-a753-f13cb10e2754"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.566418 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwjgf\" (UniqueName: \"kubernetes.io/projected/d9d66796-ff80-4c4b-a753-f13cb10e2754-kube-api-access-gwjgf\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.566458 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.597676 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-config" (OuterVolumeSpecName: "config") pod "d9d66796-ff80-4c4b-a753-f13cb10e2754" (UID: "d9d66796-ff80-4c4b-a753-f13cb10e2754"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.607890 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d9d66796-ff80-4c4b-a753-f13cb10e2754" (UID: "d9d66796-ff80-4c4b-a753-f13cb10e2754"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.622853 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d9d66796-ff80-4c4b-a753-f13cb10e2754" (UID: "d9d66796-ff80-4c4b-a753-f13cb10e2754"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.624912 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d9d66796-ff80-4c4b-a753-f13cb10e2754" (UID: "d9d66796-ff80-4c4b-a753-f13cb10e2754"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.670328 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.670382 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.670736 4771 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.670757 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9d66796-ff80-4c4b-a753-f13cb10e2754-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.896157 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.987451 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vj26\" (UniqueName: \"kubernetes.io/projected/325b9e47-f103-41d4-ad1f-f9fb2eecde45-kube-api-access-9vj26\") pod \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\" (UID: \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\") " Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.987559 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/325b9e47-f103-41d4-ad1f-f9fb2eecde45-config-volume\") pod \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\" (UID: \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\") " Oct 02 10:00:03 crc kubenswrapper[4771]: I1002 10:00:03.987730 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/325b9e47-f103-41d4-ad1f-f9fb2eecde45-secret-volume\") pod \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\" (UID: \"325b9e47-f103-41d4-ad1f-f9fb2eecde45\") " Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.004490 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/325b9e47-f103-41d4-ad1f-f9fb2eecde45-config-volume" (OuterVolumeSpecName: "config-volume") pod "325b9e47-f103-41d4-ad1f-f9fb2eecde45" (UID: "325b9e47-f103-41d4-ad1f-f9fb2eecde45"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.024491 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/325b9e47-f103-41d4-ad1f-f9fb2eecde45-kube-api-access-9vj26" (OuterVolumeSpecName: "kube-api-access-9vj26") pod "325b9e47-f103-41d4-ad1f-f9fb2eecde45" (UID: "325b9e47-f103-41d4-ad1f-f9fb2eecde45"). InnerVolumeSpecName "kube-api-access-9vj26". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.052410 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/325b9e47-f103-41d4-ad1f-f9fb2eecde45-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "325b9e47-f103-41d4-ad1f-f9fb2eecde45" (UID: "325b9e47-f103-41d4-ad1f-f9fb2eecde45"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.098571 4771 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/325b9e47-f103-41d4-ad1f-f9fb2eecde45-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.098649 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vj26\" (UniqueName: \"kubernetes.io/projected/325b9e47-f103-41d4-ad1f-f9fb2eecde45-kube-api-access-9vj26\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.098667 4771 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/325b9e47-f103-41d4-ad1f-f9fb2eecde45-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.240215 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cjqrx" event={"ID":"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a","Type":"ContainerStarted","Data":"fbefaa99f273d9c68cd7432a508e995273b63587f1ce7684d16c9863d6a16585"} Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.252038 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" event={"ID":"325b9e47-f103-41d4-ad1f-f9fb2eecde45","Type":"ContainerDied","Data":"5f081a1501b0b695ef271e7492ff39c8712c434a3f9b07f930adb6aa05d0ff72"} Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.252339 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f081a1501b0b695ef271e7492ff39c8712c434a3f9b07f930adb6aa05d0ff72" Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.252474 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7" Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.268157 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" event={"ID":"d9d66796-ff80-4c4b-a753-f13cb10e2754","Type":"ContainerDied","Data":"8b3f9bbe85b0a45f6659e13d7f15c56cd09079547f2716e6c34957fa689e26cc"} Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.268233 4771 scope.go:117] "RemoveContainer" containerID="db31a9ebdb450dfb16c11ed4a92b24c31374708125c0aece21d9333d8d67f7cb" Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.268408 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-kjwbj" Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.269335 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-x9s6s"] Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.276289 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-cjqrx" podStartSLOduration=3.707999505 podStartE2EDuration="58.276263942s" podCreationTimestamp="2025-10-02 09:59:06 +0000 UTC" firstStartedPulling="2025-10-02 09:59:08.359214746 +0000 UTC m=+1336.006899813" lastFinishedPulling="2025-10-02 10:00:02.927479183 +0000 UTC m=+1390.575164250" observedRunningTime="2025-10-02 10:00:04.274027834 +0000 UTC m=+1391.921712911" watchObservedRunningTime="2025-10-02 10:00:04.276263942 +0000 UTC m=+1391.923949029" Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.315183 4771 scope.go:117] "RemoveContainer" containerID="489d8258995a5b34243e22f381ff037037aed40983bda64927b307c8efc16d0a" Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.330650 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-kjwbj"] Oct 02 10:00:04 crc kubenswrapper[4771]: I1002 10:00:04.332537 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-kjwbj"] Oct 02 10:00:05 crc kubenswrapper[4771]: I1002 10:00:05.300049 4771 generic.go:334] "Generic (PLEG): container finished" podID="5dcb591a-9c0b-44ce-a0f0-db26883d5479" containerID="ee183dcedf71d15317c762f546597ed9e5af9f64a11e3eb17ac19a77251681af" exitCode=0 Oct 02 10:00:05 crc kubenswrapper[4771]: I1002 10:00:05.300957 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" event={"ID":"5dcb591a-9c0b-44ce-a0f0-db26883d5479","Type":"ContainerDied","Data":"ee183dcedf71d15317c762f546597ed9e5af9f64a11e3eb17ac19a77251681af"} Oct 02 10:00:05 crc kubenswrapper[4771]: I1002 10:00:05.300994 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" event={"ID":"5dcb591a-9c0b-44ce-a0f0-db26883d5479","Type":"ContainerStarted","Data":"10294a001b0998054be9d54032b00a30095c2b7fc6f8043aeaedb08573b2eac3"} Oct 02 10:00:05 crc kubenswrapper[4771]: I1002 10:00:05.700400 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9d66796-ff80-4c4b-a753-f13cb10e2754" path="/var/lib/kubelet/pods/d9d66796-ff80-4c4b-a753-f13cb10e2754/volumes" Oct 02 10:00:06 crc kubenswrapper[4771]: I1002 10:00:06.320844 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" event={"ID":"5dcb591a-9c0b-44ce-a0f0-db26883d5479","Type":"ContainerStarted","Data":"1148de2b7f720897d7f55a895c6aeef8a59fa8f1b2ebf644f4e153c9e9652fff"} Oct 02 10:00:06 crc kubenswrapper[4771]: I1002 10:00:06.321218 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:06 crc kubenswrapper[4771]: I1002 10:00:06.347004 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" podStartSLOduration=4.346974872 podStartE2EDuration="4.346974872s" podCreationTimestamp="2025-10-02 10:00:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:06.343500513 +0000 UTC m=+1393.991185600" watchObservedRunningTime="2025-10-02 10:00:06.346974872 +0000 UTC m=+1393.994659939" Oct 02 10:00:07 crc kubenswrapper[4771]: I1002 10:00:07.339306 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6caafe76-3bdf-423b-abb6-c2057f835d38","Type":"ContainerStarted","Data":"ce4404e24ea88a130fd7d6dc155aa3fc90a4425f21df6ccd8efe1954ca11d8e1"} Oct 02 10:00:07 crc kubenswrapper[4771]: I1002 10:00:07.339709 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6caafe76-3bdf-423b-abb6-c2057f835d38","Type":"ContainerStarted","Data":"c109ef3cebd331b9cdfd134c6c7a2caaf1e11f7b08a4eac49afa3f33f90c2167"} Oct 02 10:00:07 crc kubenswrapper[4771]: I1002 10:00:07.392774 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=37.392706642 podStartE2EDuration="37.392706642s" podCreationTimestamp="2025-10-02 09:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:07.38053781 +0000 UTC m=+1395.028222887" watchObservedRunningTime="2025-10-02 10:00:07.392706642 +0000 UTC m=+1395.040391709" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.359443 4771 generic.go:334] "Generic (PLEG): container finished" podID="9f0f7ed9-b19d-41e3-9eb3-d6431979e26a" containerID="fbefaa99f273d9c68cd7432a508e995273b63587f1ce7684d16c9863d6a16585" exitCode=0 Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.359528 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cjqrx" event={"ID":"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a","Type":"ContainerDied","Data":"fbefaa99f273d9c68cd7432a508e995273b63587f1ce7684d16c9863d6a16585"} Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.669146 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j4c5j"] Oct 02 10:00:09 crc kubenswrapper[4771]: E1002 10:00:09.676676 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="325b9e47-f103-41d4-ad1f-f9fb2eecde45" containerName="collect-profiles" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.676728 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="325b9e47-f103-41d4-ad1f-f9fb2eecde45" containerName="collect-profiles" Oct 02 10:00:09 crc kubenswrapper[4771]: E1002 10:00:09.676780 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d66796-ff80-4c4b-a753-f13cb10e2754" containerName="dnsmasq-dns" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.676791 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d66796-ff80-4c4b-a753-f13cb10e2754" containerName="dnsmasq-dns" Oct 02 10:00:09 crc kubenswrapper[4771]: E1002 10:00:09.676828 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d66796-ff80-4c4b-a753-f13cb10e2754" containerName="init" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.676840 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d66796-ff80-4c4b-a753-f13cb10e2754" containerName="init" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.677263 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d66796-ff80-4c4b-a753-f13cb10e2754" containerName="dnsmasq-dns" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.677291 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="325b9e47-f103-41d4-ad1f-f9fb2eecde45" containerName="collect-profiles" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.680661 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.741527 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j4c5j"] Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.774529 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3184235b-9379-4b54-803d-dbb52582e06b-utilities\") pod \"redhat-operators-j4c5j\" (UID: \"3184235b-9379-4b54-803d-dbb52582e06b\") " pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.774764 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3184235b-9379-4b54-803d-dbb52582e06b-catalog-content\") pod \"redhat-operators-j4c5j\" (UID: \"3184235b-9379-4b54-803d-dbb52582e06b\") " pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.774794 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfm2f\" (UniqueName: \"kubernetes.io/projected/3184235b-9379-4b54-803d-dbb52582e06b-kube-api-access-tfm2f\") pod \"redhat-operators-j4c5j\" (UID: \"3184235b-9379-4b54-803d-dbb52582e06b\") " pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.876911 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3184235b-9379-4b54-803d-dbb52582e06b-utilities\") pod \"redhat-operators-j4c5j\" (UID: \"3184235b-9379-4b54-803d-dbb52582e06b\") " pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.877337 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3184235b-9379-4b54-803d-dbb52582e06b-catalog-content\") pod \"redhat-operators-j4c5j\" (UID: \"3184235b-9379-4b54-803d-dbb52582e06b\") " pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.877383 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfm2f\" (UniqueName: \"kubernetes.io/projected/3184235b-9379-4b54-803d-dbb52582e06b-kube-api-access-tfm2f\") pod \"redhat-operators-j4c5j\" (UID: \"3184235b-9379-4b54-803d-dbb52582e06b\") " pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.877414 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3184235b-9379-4b54-803d-dbb52582e06b-utilities\") pod \"redhat-operators-j4c5j\" (UID: \"3184235b-9379-4b54-803d-dbb52582e06b\") " pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.878044 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3184235b-9379-4b54-803d-dbb52582e06b-catalog-content\") pod \"redhat-operators-j4c5j\" (UID: \"3184235b-9379-4b54-803d-dbb52582e06b\") " pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:00:09 crc kubenswrapper[4771]: I1002 10:00:09.897526 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfm2f\" (UniqueName: \"kubernetes.io/projected/3184235b-9379-4b54-803d-dbb52582e06b-kube-api-access-tfm2f\") pod \"redhat-operators-j4c5j\" (UID: \"3184235b-9379-4b54-803d-dbb52582e06b\") " pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:00:10 crc kubenswrapper[4771]: I1002 10:00:10.025620 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:00:10 crc kubenswrapper[4771]: I1002 10:00:10.562366 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j4c5j"] Oct 02 10:00:10 crc kubenswrapper[4771]: I1002 10:00:10.890989 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 02 10:00:10 crc kubenswrapper[4771]: I1002 10:00:10.901106 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cjqrx" Oct 02 10:00:10 crc kubenswrapper[4771]: I1002 10:00:10.915956 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-config-data\") pod \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\" (UID: \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\") " Oct 02 10:00:10 crc kubenswrapper[4771]: I1002 10:00:10.916207 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbfbp\" (UniqueName: \"kubernetes.io/projected/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-kube-api-access-mbfbp\") pod \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\" (UID: \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\") " Oct 02 10:00:10 crc kubenswrapper[4771]: I1002 10:00:10.916514 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-combined-ca-bundle\") pod \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\" (UID: \"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a\") " Oct 02 10:00:10 crc kubenswrapper[4771]: I1002 10:00:10.930110 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-kube-api-access-mbfbp" (OuterVolumeSpecName: "kube-api-access-mbfbp") pod "9f0f7ed9-b19d-41e3-9eb3-d6431979e26a" (UID: "9f0f7ed9-b19d-41e3-9eb3-d6431979e26a"). InnerVolumeSpecName "kube-api-access-mbfbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:10 crc kubenswrapper[4771]: I1002 10:00:10.968754 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f0f7ed9-b19d-41e3-9eb3-d6431979e26a" (UID: "9f0f7ed9-b19d-41e3-9eb3-d6431979e26a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.031451 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.031479 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbfbp\" (UniqueName: \"kubernetes.io/projected/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-kube-api-access-mbfbp\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.065910 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-config-data" (OuterVolumeSpecName: "config-data") pod "9f0f7ed9-b19d-41e3-9eb3-d6431979e26a" (UID: "9f0f7ed9-b19d-41e3-9eb3-d6431979e26a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.133687 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.404282 4771 generic.go:334] "Generic (PLEG): container finished" podID="3184235b-9379-4b54-803d-dbb52582e06b" containerID="4b23ae1a19066feda4479e04b20e05bd58683722fdfcfb9179f7f69c739fb215" exitCode=0 Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.404360 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4c5j" event={"ID":"3184235b-9379-4b54-803d-dbb52582e06b","Type":"ContainerDied","Data":"4b23ae1a19066feda4479e04b20e05bd58683722fdfcfb9179f7f69c739fb215"} Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.404398 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4c5j" event={"ID":"3184235b-9379-4b54-803d-dbb52582e06b","Type":"ContainerStarted","Data":"8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f"} Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.422861 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cjqrx" event={"ID":"9f0f7ed9-b19d-41e3-9eb3-d6431979e26a","Type":"ContainerDied","Data":"fc3175fb2bc3c0d7ba4850ec237eccf26391a2694749b11739348a4436236e84"} Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.422922 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc3175fb2bc3c0d7ba4850ec237eccf26391a2694749b11739348a4436236e84" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.423001 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cjqrx" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.845834 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-bm9ln"] Oct 02 10:00:11 crc kubenswrapper[4771]: E1002 10:00:11.846989 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0f7ed9-b19d-41e3-9eb3-d6431979e26a" containerName="keystone-db-sync" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.847014 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0f7ed9-b19d-41e3-9eb3-d6431979e26a" containerName="keystone-db-sync" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.847318 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0f7ed9-b19d-41e3-9eb3-d6431979e26a" containerName="keystone-db-sync" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.851869 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.860829 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.861077 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.861290 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.861466 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8kg7s" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.863146 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-scripts\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.863315 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-combined-ca-bundle\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.863362 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58qnb\" (UniqueName: \"kubernetes.io/projected/ec499dd8-fc17-46bd-8771-eb78b728da89-kube-api-access-58qnb\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.863439 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-fernet-keys\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.863483 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-credential-keys\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.863533 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-config-data\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.880710 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-x9s6s"] Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.881033 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" podUID="5dcb591a-9c0b-44ce-a0f0-db26883d5479" containerName="dnsmasq-dns" containerID="cri-o://1148de2b7f720897d7f55a895c6aeef8a59fa8f1b2ebf644f4e153c9e9652fff" gracePeriod=10 Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.885909 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.907890 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bm9ln"] Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.972282 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-fernet-keys\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.972344 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-credential-keys\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.972377 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-config-data\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.972492 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-scripts\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.986017 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-combined-ca-bundle\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:11 crc kubenswrapper[4771]: I1002 10:00:11.986160 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58qnb\" (UniqueName: \"kubernetes.io/projected/ec499dd8-fc17-46bd-8771-eb78b728da89-kube-api-access-58qnb\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.006864 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-fernet-keys\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.015417 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-config-data\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.017819 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-credential-keys\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.023379 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-combined-ca-bundle\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.035424 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-scripts\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.042819 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58qnb\" (UniqueName: \"kubernetes.io/projected/ec499dd8-fc17-46bd-8771-eb78b728da89-kube-api-access-58qnb\") pod \"keystone-bootstrap-bm9ln\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.066054 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-vdt6g"] Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.085410 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.180152 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-vdt6g"] Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.193497 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-config\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.193576 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.193674 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.193757 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcng4\" (UniqueName: \"kubernetes.io/projected/851bcee9-3b3a-46e9-a950-fcf6cee435ba-kube-api-access-zcng4\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.193779 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-dns-svc\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.195065 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.200538 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.201570 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-dbght"] Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.222193 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-dbght" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.240307 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-4fps8" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.240546 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.342728 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-pkkwb"] Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.369072 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.380713 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.381014 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jqk64" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.386229 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-config\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.386308 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.386412 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.386454 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042a2fb4-504c-4448-922b-48db5507e964-combined-ca-bundle\") pod \"heat-db-sync-dbght\" (UID: \"042a2fb4-504c-4448-922b-48db5507e964\") " pod="openstack/heat-db-sync-dbght" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.386483 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcng4\" (UniqueName: \"kubernetes.io/projected/851bcee9-3b3a-46e9-a950-fcf6cee435ba-kube-api-access-zcng4\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.386509 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d56n\" (UniqueName: \"kubernetes.io/projected/042a2fb4-504c-4448-922b-48db5507e964-kube-api-access-8d56n\") pod \"heat-db-sync-dbght\" (UID: \"042a2fb4-504c-4448-922b-48db5507e964\") " pod="openstack/heat-db-sync-dbght" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.386541 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-dns-svc\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.386598 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.386670 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042a2fb4-504c-4448-922b-48db5507e964-config-data\") pod \"heat-db-sync-dbght\" (UID: \"042a2fb4-504c-4448-922b-48db5507e964\") " pod="openstack/heat-db-sync-dbght" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.389960 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.391372 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.392835 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-dns-svc\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.393458 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.393714 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.402744 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-config\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.494729 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042a2fb4-504c-4448-922b-48db5507e964-combined-ca-bundle\") pod \"heat-db-sync-dbght\" (UID: \"042a2fb4-504c-4448-922b-48db5507e964\") " pod="openstack/heat-db-sync-dbght" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.494808 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d56n\" (UniqueName: \"kubernetes.io/projected/042a2fb4-504c-4448-922b-48db5507e964-kube-api-access-8d56n\") pod \"heat-db-sync-dbght\" (UID: \"042a2fb4-504c-4448-922b-48db5507e964\") " pod="openstack/heat-db-sync-dbght" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.494860 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-combined-ca-bundle\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.494901 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twr2n\" (UniqueName: \"kubernetes.io/projected/2a276f13-9f48-40ae-8445-ce376fce62a1-kube-api-access-twr2n\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.494936 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a276f13-9f48-40ae-8445-ce376fce62a1-etc-machine-id\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.494974 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042a2fb4-504c-4448-922b-48db5507e964-config-data\") pod \"heat-db-sync-dbght\" (UID: \"042a2fb4-504c-4448-922b-48db5507e964\") " pod="openstack/heat-db-sync-dbght" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.495005 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-scripts\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.495061 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-config-data\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.495098 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-db-sync-config-data\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.539111 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042a2fb4-504c-4448-922b-48db5507e964-combined-ca-bundle\") pod \"heat-db-sync-dbght\" (UID: \"042a2fb4-504c-4448-922b-48db5507e964\") " pod="openstack/heat-db-sync-dbght" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.541090 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcng4\" (UniqueName: \"kubernetes.io/projected/851bcee9-3b3a-46e9-a950-fcf6cee435ba-kube-api-access-zcng4\") pod \"dnsmasq-dns-847c4cc679-vdt6g\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.544544 4771 generic.go:334] "Generic (PLEG): container finished" podID="5dcb591a-9c0b-44ce-a0f0-db26883d5479" containerID="1148de2b7f720897d7f55a895c6aeef8a59fa8f1b2ebf644f4e153c9e9652fff" exitCode=0 Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.544606 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" event={"ID":"5dcb591a-9c0b-44ce-a0f0-db26883d5479","Type":"ContainerDied","Data":"1148de2b7f720897d7f55a895c6aeef8a59fa8f1b2ebf644f4e153c9e9652fff"} Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.548849 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042a2fb4-504c-4448-922b-48db5507e964-config-data\") pod \"heat-db-sync-dbght\" (UID: \"042a2fb4-504c-4448-922b-48db5507e964\") " pod="openstack/heat-db-sync-dbght" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.566731 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.587222 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-pkkwb"] Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.593165 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d56n\" (UniqueName: \"kubernetes.io/projected/042a2fb4-504c-4448-922b-48db5507e964-kube-api-access-8d56n\") pod \"heat-db-sync-dbght\" (UID: \"042a2fb4-504c-4448-922b-48db5507e964\") " pod="openstack/heat-db-sync-dbght" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.599541 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-combined-ca-bundle\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.600050 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twr2n\" (UniqueName: \"kubernetes.io/projected/2a276f13-9f48-40ae-8445-ce376fce62a1-kube-api-access-twr2n\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.600100 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a276f13-9f48-40ae-8445-ce376fce62a1-etc-machine-id\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.600215 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-scripts\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.600277 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-config-data\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.600329 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-db-sync-config-data\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.604350 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a276f13-9f48-40ae-8445-ce376fce62a1-etc-machine-id\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.605405 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-db-sync-config-data\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.611333 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-config-data\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.630881 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-dbght" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.644979 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-combined-ca-bundle\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.645010 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-scripts\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.687351 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twr2n\" (UniqueName: \"kubernetes.io/projected/2a276f13-9f48-40ae-8445-ce376fce62a1-kube-api-access-twr2n\") pod \"cinder-db-sync-pkkwb\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.710795 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-dbght"] Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.855829 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.861061 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-f725t"] Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.864923 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-f725t" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.868558 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.869379 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.869507 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nn87q" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.899614 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-5dxhb"] Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.902444 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5dxhb" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.906629 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.906837 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-lqb76" Oct 02 10:00:12 crc kubenswrapper[4771]: I1002 10:00:12.965941 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-f725t"] Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.003444 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5dxhb"] Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.033940 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975630d7-f63d-43eb-99c4-a245f4f20c24-combined-ca-bundle\") pod \"neutron-db-sync-f725t\" (UID: \"975630d7-f63d-43eb-99c4-a245f4f20c24\") " pod="openstack/neutron-db-sync-f725t" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.034044 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bc64084-015e-47d2-b20c-f241e367fe27-db-sync-config-data\") pod \"barbican-db-sync-5dxhb\" (UID: \"3bc64084-015e-47d2-b20c-f241e367fe27\") " pod="openstack/barbican-db-sync-5dxhb" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.034092 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wv5l\" (UniqueName: \"kubernetes.io/projected/3bc64084-015e-47d2-b20c-f241e367fe27-kube-api-access-9wv5l\") pod \"barbican-db-sync-5dxhb\" (UID: \"3bc64084-015e-47d2-b20c-f241e367fe27\") " pod="openstack/barbican-db-sync-5dxhb" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.034160 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/975630d7-f63d-43eb-99c4-a245f4f20c24-config\") pod \"neutron-db-sync-f725t\" (UID: \"975630d7-f63d-43eb-99c4-a245f4f20c24\") " pod="openstack/neutron-db-sync-f725t" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.034240 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vckbr\" (UniqueName: \"kubernetes.io/projected/975630d7-f63d-43eb-99c4-a245f4f20c24-kube-api-access-vckbr\") pod \"neutron-db-sync-f725t\" (UID: \"975630d7-f63d-43eb-99c4-a245f4f20c24\") " pod="openstack/neutron-db-sync-f725t" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.034316 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc64084-015e-47d2-b20c-f241e367fe27-combined-ca-bundle\") pod \"barbican-db-sync-5dxhb\" (UID: \"3bc64084-015e-47d2-b20c-f241e367fe27\") " pod="openstack/barbican-db-sync-5dxhb" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.133205 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-wg2x2"] Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.137861 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975630d7-f63d-43eb-99c4-a245f4f20c24-combined-ca-bundle\") pod \"neutron-db-sync-f725t\" (UID: \"975630d7-f63d-43eb-99c4-a245f4f20c24\") " pod="openstack/neutron-db-sync-f725t" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.137955 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bc64084-015e-47d2-b20c-f241e367fe27-db-sync-config-data\") pod \"barbican-db-sync-5dxhb\" (UID: \"3bc64084-015e-47d2-b20c-f241e367fe27\") " pod="openstack/barbican-db-sync-5dxhb" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.137994 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wv5l\" (UniqueName: \"kubernetes.io/projected/3bc64084-015e-47d2-b20c-f241e367fe27-kube-api-access-9wv5l\") pod \"barbican-db-sync-5dxhb\" (UID: \"3bc64084-015e-47d2-b20c-f241e367fe27\") " pod="openstack/barbican-db-sync-5dxhb" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.138041 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/975630d7-f63d-43eb-99c4-a245f4f20c24-config\") pod \"neutron-db-sync-f725t\" (UID: \"975630d7-f63d-43eb-99c4-a245f4f20c24\") " pod="openstack/neutron-db-sync-f725t" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.138120 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vckbr\" (UniqueName: \"kubernetes.io/projected/975630d7-f63d-43eb-99c4-a245f4f20c24-kube-api-access-vckbr\") pod \"neutron-db-sync-f725t\" (UID: \"975630d7-f63d-43eb-99c4-a245f4f20c24\") " pod="openstack/neutron-db-sync-f725t" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.138219 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc64084-015e-47d2-b20c-f241e367fe27-combined-ca-bundle\") pod \"barbican-db-sync-5dxhb\" (UID: \"3bc64084-015e-47d2-b20c-f241e367fe27\") " pod="openstack/barbican-db-sync-5dxhb" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.140801 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.156886 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc64084-015e-47d2-b20c-f241e367fe27-combined-ca-bundle\") pod \"barbican-db-sync-5dxhb\" (UID: \"3bc64084-015e-47d2-b20c-f241e367fe27\") " pod="openstack/barbican-db-sync-5dxhb" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.215481 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/975630d7-f63d-43eb-99c4-a245f4f20c24-config\") pod \"neutron-db-sync-f725t\" (UID: \"975630d7-f63d-43eb-99c4-a245f4f20c24\") " pod="openstack/neutron-db-sync-f725t" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.218565 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-fmxrj" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.223727 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.224937 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.229636 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975630d7-f63d-43eb-99c4-a245f4f20c24-combined-ca-bundle\") pod \"neutron-db-sync-f725t\" (UID: \"975630d7-f63d-43eb-99c4-a245f4f20c24\") " pod="openstack/neutron-db-sync-f725t" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.230019 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-wg2x2"] Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.240717 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bc64084-015e-47d2-b20c-f241e367fe27-db-sync-config-data\") pod \"barbican-db-sync-5dxhb\" (UID: \"3bc64084-015e-47d2-b20c-f241e367fe27\") " pod="openstack/barbican-db-sync-5dxhb" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.257627 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-logs\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.263649 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-scripts\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.263917 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khdp2\" (UniqueName: \"kubernetes.io/projected/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-kube-api-access-khdp2\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.264015 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-config-data\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.265365 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-combined-ca-bundle\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.268659 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wv5l\" (UniqueName: \"kubernetes.io/projected/3bc64084-015e-47d2-b20c-f241e367fe27-kube-api-access-9wv5l\") pod \"barbican-db-sync-5dxhb\" (UID: \"3bc64084-015e-47d2-b20c-f241e367fe27\") " pod="openstack/barbican-db-sync-5dxhb" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.268818 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-vdt6g"] Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.283036 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5dxhb" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.294374 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-qr4df"] Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.298678 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.302521 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vckbr\" (UniqueName: \"kubernetes.io/projected/975630d7-f63d-43eb-99c4-a245f4f20c24-kube-api-access-vckbr\") pod \"neutron-db-sync-f725t\" (UID: \"975630d7-f63d-43eb-99c4-a245f4f20c24\") " pod="openstack/neutron-db-sync-f725t" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.312739 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-qr4df"] Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.370793 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.373621 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-combined-ca-bundle\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.373785 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-logs\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.373875 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-scripts\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.373920 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khdp2\" (UniqueName: \"kubernetes.io/projected/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-kube-api-access-khdp2\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.373955 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-config-data\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.389629 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-logs\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.392041 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.395394 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.400018 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-config-data\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.400705 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.400943 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lnfgz" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.401087 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.402416 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.403873 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-combined-ca-bundle\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.404928 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-scripts\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.425099 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khdp2\" (UniqueName: \"kubernetes.io/projected/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-kube-api-access-khdp2\") pod \"placement-db-sync-wg2x2\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.480835 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.480902 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.480995 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf9xw\" (UniqueName: \"kubernetes.io/projected/eb11e684-5673-479e-873b-d61930fe2d52-kube-api-access-kf9xw\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.481017 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.481148 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.481196 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-config\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.481305 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.483841 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.490164 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.490513 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.512638 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.550065 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-f725t" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.583068 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.592468 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fea444ce-0980-4dd4-a4d4-f391e79f6a05-logs\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.592700 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf9xw\" (UniqueName: \"kubernetes.io/projected/eb11e684-5673-479e-873b-d61930fe2d52-kube-api-access-kf9xw\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.592724 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.592764 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-scripts\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.592822 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-config-data\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.592907 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.592934 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgmhn\" (UniqueName: \"kubernetes.io/projected/109572b4-1dd0-4e0e-a02f-f307c07ec065-kube-api-access-dgmhn\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.592982 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/109572b4-1dd0-4e0e-a02f-f307c07ec065-logs\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.593035 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.593064 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-config-data\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.593109 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.593192 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-scripts\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.593270 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.593315 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.593350 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.593386 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-config\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.593451 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6bxz\" (UniqueName: \"kubernetes.io/projected/fea444ce-0980-4dd4-a4d4-f391e79f6a05-kube-api-access-r6bxz\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.593494 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.593573 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.593599 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/109572b4-1dd0-4e0e-a02f-f307c07ec065-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.593628 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fea444ce-0980-4dd4-a4d4-f391e79f6a05-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.595557 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.599169 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.604222 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.605986 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.606591 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-config\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.635041 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.651080 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf9xw\" (UniqueName: \"kubernetes.io/projected/eb11e684-5673-479e-873b-d61930fe2d52-kube-api-access-kf9xw\") pod \"dnsmasq-dns-785d8bcb8c-qr4df\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.704306 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-dns-svc\") pod \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.704478 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-config\") pod \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.704594 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-ovsdbserver-sb\") pod \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.704678 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdmk7\" (UniqueName: \"kubernetes.io/projected/5dcb591a-9c0b-44ce-a0f0-db26883d5479-kube-api-access-kdmk7\") pod \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.704729 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-ovsdbserver-nb\") pod \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.704751 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-dns-swift-storage-0\") pod \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705076 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fea444ce-0980-4dd4-a4d4-f391e79f6a05-logs\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705202 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-scripts\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705241 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-config-data\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705294 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705322 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgmhn\" (UniqueName: \"kubernetes.io/projected/109572b4-1dd0-4e0e-a02f-f307c07ec065-kube-api-access-dgmhn\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705362 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/109572b4-1dd0-4e0e-a02f-f307c07ec065-logs\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705405 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705437 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-config-data\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705472 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705534 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-scripts\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705573 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705598 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705652 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6bxz\" (UniqueName: \"kubernetes.io/projected/fea444ce-0980-4dd4-a4d4-f391e79f6a05-kube-api-access-r6bxz\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705691 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/109572b4-1dd0-4e0e-a02f-f307c07ec065-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705739 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fea444ce-0980-4dd4-a4d4-f391e79f6a05-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.705776 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.707967 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wg2x2" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.713939 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.716479 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.716777 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/109572b4-1dd0-4e0e-a02f-f307c07ec065-logs\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.717173 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/109572b4-1dd0-4e0e-a02f-f307c07ec065-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.719845 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fea444ce-0980-4dd4-a4d4-f391e79f6a05-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.726946 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-scripts\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.730607 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fea444ce-0980-4dd4-a4d4-f391e79f6a05-logs\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.787252 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.793151 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.795924 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-config-data\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.798008 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.802954 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgmhn\" (UniqueName: \"kubernetes.io/projected/109572b4-1dd0-4e0e-a02f-f307c07ec065-kube-api-access-dgmhn\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.806716 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dcb591a-9c0b-44ce-a0f0-db26883d5479-kube-api-access-kdmk7" (OuterVolumeSpecName: "kube-api-access-kdmk7") pod "5dcb591a-9c0b-44ce-a0f0-db26883d5479" (UID: "5dcb591a-9c0b-44ce-a0f0-db26883d5479"). InnerVolumeSpecName "kube-api-access-kdmk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.812088 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.822222 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdmk7\" (UniqueName: \"kubernetes.io/projected/5dcb591a-9c0b-44ce-a0f0-db26883d5479-kube-api-access-kdmk7\") pod \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\" (UID: \"5dcb591a-9c0b-44ce-a0f0-db26883d5479\") " Oct 02 10:00:13 crc kubenswrapper[4771]: W1002 10:00:13.824934 4771 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/5dcb591a-9c0b-44ce-a0f0-db26883d5479/volumes/kubernetes.io~projected/kube-api-access-kdmk7 Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.825051 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dcb591a-9c0b-44ce-a0f0-db26883d5479-kube-api-access-kdmk7" (OuterVolumeSpecName: "kube-api-access-kdmk7") pod "5dcb591a-9c0b-44ce-a0f0-db26883d5479" (UID: "5dcb591a-9c0b-44ce-a0f0-db26883d5479"). InnerVolumeSpecName "kube-api-access-kdmk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.911685 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.911705 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-scripts\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:13 crc kubenswrapper[4771]: I1002 10:00:13.969344 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdmk7\" (UniqueName: \"kubernetes.io/projected/5dcb591a-9c0b-44ce-a0f0-db26883d5479-kube-api-access-kdmk7\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.029096 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-config-data\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.029776 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6bxz\" (UniqueName: \"kubernetes.io/projected/fea444ce-0980-4dd4-a4d4-f391e79f6a05-kube-api-access-r6bxz\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.119986 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:00:14 crc kubenswrapper[4771]: E1002 10:00:14.120673 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dcb591a-9c0b-44ce-a0f0-db26883d5479" containerName="init" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.120694 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dcb591a-9c0b-44ce-a0f0-db26883d5479" containerName="init" Oct 02 10:00:14 crc kubenswrapper[4771]: E1002 10:00:14.120729 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dcb591a-9c0b-44ce-a0f0-db26883d5479" containerName="dnsmasq-dns" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.120738 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dcb591a-9c0b-44ce-a0f0-db26883d5479" containerName="dnsmasq-dns" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.121022 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dcb591a-9c0b-44ce-a0f0-db26883d5479" containerName="dnsmasq-dns" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.135815 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.150170 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.164825 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bm9ln"] Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.188055 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.188371 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.243419 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48fe997b-eed7-4191-9f63-ecb40112cf50-run-httpd\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.243818 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48fe997b-eed7-4191-9f63-ecb40112cf50-log-httpd\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.272056 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-scripts\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.272866 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jzxj\" (UniqueName: \"kubernetes.io/projected/48fe997b-eed7-4191-9f63-ecb40112cf50-kube-api-access-9jzxj\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.273113 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.273726 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-config-data\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.274044 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.307778 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.382779 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48fe997b-eed7-4191-9f63-ecb40112cf50-log-httpd\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.382943 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-scripts\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.382972 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jzxj\" (UniqueName: \"kubernetes.io/projected/48fe997b-eed7-4191-9f63-ecb40112cf50-kube-api-access-9jzxj\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.383000 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.383034 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-config-data\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.383063 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.383101 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48fe997b-eed7-4191-9f63-ecb40112cf50-run-httpd\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.383671 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48fe997b-eed7-4191-9f63-ecb40112cf50-run-httpd\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.383875 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48fe997b-eed7-4191-9f63-ecb40112cf50-log-httpd\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.411457 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-config-data\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.415496 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.416212 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.430545 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.467282 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-scripts\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.486685 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.486953 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jzxj\" (UniqueName: \"kubernetes.io/projected/48fe997b-eed7-4191-9f63-ecb40112cf50-kube-api-access-9jzxj\") pod \"ceilometer-0\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.515093 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5dcb591a-9c0b-44ce-a0f0-db26883d5479" (UID: "5dcb591a-9c0b-44ce-a0f0-db26883d5479"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.589557 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.598516 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-config" (OuterVolumeSpecName: "config") pod "5dcb591a-9c0b-44ce-a0f0-db26883d5479" (UID: "5dcb591a-9c0b-44ce-a0f0-db26883d5479"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.613972 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5dcb591a-9c0b-44ce-a0f0-db26883d5479" (UID: "5dcb591a-9c0b-44ce-a0f0-db26883d5479"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.669930 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5dcb591a-9c0b-44ce-a0f0-db26883d5479" (UID: "5dcb591a-9c0b-44ce-a0f0-db26883d5479"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.682916 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5dcb591a-9c0b-44ce-a0f0-db26883d5479" (UID: "5dcb591a-9c0b-44ce-a0f0-db26883d5479"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.684614 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.696927 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.696966 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.696978 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.696993 4771 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5dcb591a-9c0b-44ce-a0f0-db26883d5479-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.702900 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4c5j" event={"ID":"3184235b-9379-4b54-803d-dbb52582e06b","Type":"ContainerStarted","Data":"c4feb8e59691a866aa029f4ac85d391f25603c66aeefe0bae491495d70f46b87"} Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.717885 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bm9ln" event={"ID":"ec499dd8-fc17-46bd-8771-eb78b728da89","Type":"ContainerStarted","Data":"543aeeb78a6ffd39a616fe47a54200a595f2ef6fcbbfaebbb7808d5f7026582d"} Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.744752 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" event={"ID":"5dcb591a-9c0b-44ce-a0f0-db26883d5479","Type":"ContainerDied","Data":"10294a001b0998054be9d54032b00a30095c2b7fc6f8043aeaedb08573b2eac3"} Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.744820 4771 scope.go:117] "RemoveContainer" containerID="1148de2b7f720897d7f55a895c6aeef8a59fa8f1b2ebf644f4e153c9e9652fff" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.745214 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.758722 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-vdt6g"] Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.764468 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.837446 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-dbght"] Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.850276 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-x9s6s"] Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.868746 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-x9s6s"] Oct 02 10:00:14 crc kubenswrapper[4771]: I1002 10:00:14.983660 4771 scope.go:117] "RemoveContainer" containerID="ee183dcedf71d15317c762f546597ed9e5af9f64a11e3eb17ac19a77251681af" Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.093246 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-pkkwb"] Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.638553 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.716787 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dcb591a-9c0b-44ce-a0f0-db26883d5479" path="/var/lib/kubelet/pods/5dcb591a-9c0b-44ce-a0f0-db26883d5479/volumes" Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.720036 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5dxhb"] Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.745606 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-f725t"] Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.764027 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-wg2x2"] Oct 02 10:00:15 crc kubenswrapper[4771]: W1002 10:00:15.789329 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7e5fca5_60e8_4b3d_b6bb_73f59b2ecab5.slice/crio-ab59f9c82e7bce4474b8df1019c778ed326d14033177dc57bc9d0da16830f8f0 WatchSource:0}: Error finding container ab59f9c82e7bce4474b8df1019c778ed326d14033177dc57bc9d0da16830f8f0: Status 404 returned error can't find the container with id ab59f9c82e7bce4474b8df1019c778ed326d14033177dc57bc9d0da16830f8f0 Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.830602 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bm9ln" event={"ID":"ec499dd8-fc17-46bd-8771-eb78b728da89","Type":"ContainerStarted","Data":"64a69a71b8ee1037e5918d2efd2826cac3d09f55d8d7bc0c97a2614cb0f260b3"} Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.841105 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.849400 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pkkwb" event={"ID":"2a276f13-9f48-40ae-8445-ce376fce62a1","Type":"ContainerStarted","Data":"c7968dd590360b1c5058c83fad033a60ccc4433551e1bc26ee86520fb23e6fec"} Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.856626 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-bm9ln" podStartSLOduration=4.856601341 podStartE2EDuration="4.856601341s" podCreationTimestamp="2025-10-02 10:00:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:15.853006129 +0000 UTC m=+1403.500691186" watchObservedRunningTime="2025-10-02 10:00:15.856601341 +0000 UTC m=+1403.504286408" Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.869424 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-dbght" event={"ID":"042a2fb4-504c-4448-922b-48db5507e964","Type":"ContainerStarted","Data":"8981d7e4955cdf2abed1091372d0858fa931528cd3cd9192d1f96a9fa2282b82"} Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.886160 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" event={"ID":"851bcee9-3b3a-46e9-a950-fcf6cee435ba","Type":"ContainerStarted","Data":"62b2b48f464692140961535ec4384440f6747b6fc2a284b44f0e498b554dd91a"} Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.892819 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.896986 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-f725t" event={"ID":"975630d7-f63d-43eb-99c4-a245f4f20c24","Type":"ContainerStarted","Data":"49cb0c97d1134b45170e0b252cc621b8723ed6162dd48c90c356e511fd4df31e"} Oct 02 10:00:15 crc kubenswrapper[4771]: I1002 10:00:15.910864 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 02 10:00:16 crc kubenswrapper[4771]: I1002 10:00:16.224808 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-qr4df"] Oct 02 10:00:16 crc kubenswrapper[4771]: I1002 10:00:16.423408 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:00:16 crc kubenswrapper[4771]: W1002 10:00:16.516805 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48fe997b_eed7_4191_9f63_ecb40112cf50.slice/crio-a8b6e5a6541427a6b5651096dd99c42304ee54dd6f7da92fa12b9589e3b5e705 WatchSource:0}: Error finding container a8b6e5a6541427a6b5651096dd99c42304ee54dd6f7da92fa12b9589e3b5e705: Status 404 returned error can't find the container with id a8b6e5a6541427a6b5651096dd99c42304ee54dd6f7da92fa12b9589e3b5e705 Oct 02 10:00:16 crc kubenswrapper[4771]: I1002 10:00:16.521467 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:00:16 crc kubenswrapper[4771]: I1002 10:00:16.601277 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:00:16 crc kubenswrapper[4771]: I1002 10:00:16.636055 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:00:16 crc kubenswrapper[4771]: W1002 10:00:16.647322 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod109572b4_1dd0_4e0e_a02f_f307c07ec065.slice/crio-e4659721bfcda964595def3f51e014567bf2e9b2a437a03b749c4a25c8992d63 WatchSource:0}: Error finding container e4659721bfcda964595def3f51e014567bf2e9b2a437a03b749c4a25c8992d63: Status 404 returned error can't find the container with id e4659721bfcda964595def3f51e014567bf2e9b2a437a03b749c4a25c8992d63 Oct 02 10:00:16 crc kubenswrapper[4771]: I1002 10:00:16.949042 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fea444ce-0980-4dd4-a4d4-f391e79f6a05","Type":"ContainerStarted","Data":"a4fb157d30ed972b147d9bd1698fffce6c06106f3b96ee843636df0eb88e09e0"} Oct 02 10:00:16 crc kubenswrapper[4771]: I1002 10:00:16.969776 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"48fe997b-eed7-4191-9f63-ecb40112cf50","Type":"ContainerStarted","Data":"a8b6e5a6541427a6b5651096dd99c42304ee54dd6f7da92fa12b9589e3b5e705"} Oct 02 10:00:16 crc kubenswrapper[4771]: I1002 10:00:16.989364 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wg2x2" event={"ID":"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5","Type":"ContainerStarted","Data":"ab59f9c82e7bce4474b8df1019c778ed326d14033177dc57bc9d0da16830f8f0"} Oct 02 10:00:16 crc kubenswrapper[4771]: I1002 10:00:16.990933 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5dxhb" event={"ID":"3bc64084-015e-47d2-b20c-f241e367fe27","Type":"ContainerStarted","Data":"927fb362f2abeb29710ba042c5433d3f117f9e39673362f0c8626d4ea12253e3"} Oct 02 10:00:16 crc kubenswrapper[4771]: I1002 10:00:16.991886 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"109572b4-1dd0-4e0e-a02f-f307c07ec065","Type":"ContainerStarted","Data":"e4659721bfcda964595def3f51e014567bf2e9b2a437a03b749c4a25c8992d63"} Oct 02 10:00:16 crc kubenswrapper[4771]: I1002 10:00:16.993306 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" event={"ID":"eb11e684-5673-479e-873b-d61930fe2d52","Type":"ContainerStarted","Data":"d3ca1d08001b7497684e89f49b7e32471cca11c470a5dd03a11c310b44e9e9e5"} Oct 02 10:00:17 crc kubenswrapper[4771]: I1002 10:00:17.009955 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" event={"ID":"851bcee9-3b3a-46e9-a950-fcf6cee435ba","Type":"ContainerStarted","Data":"35409fa7746dd7c82183b971b2c77130a66c732a98d5855b35c20ba450df4b8d"} Oct 02 10:00:17 crc kubenswrapper[4771]: I1002 10:00:17.010166 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" podUID="851bcee9-3b3a-46e9-a950-fcf6cee435ba" containerName="init" containerID="cri-o://35409fa7746dd7c82183b971b2c77130a66c732a98d5855b35c20ba450df4b8d" gracePeriod=10 Oct 02 10:00:17 crc kubenswrapper[4771]: I1002 10:00:17.027956 4771 generic.go:334] "Generic (PLEG): container finished" podID="3184235b-9379-4b54-803d-dbb52582e06b" containerID="c4feb8e59691a866aa029f4ac85d391f25603c66aeefe0bae491495d70f46b87" exitCode=0 Oct 02 10:00:17 crc kubenswrapper[4771]: I1002 10:00:17.029813 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4c5j" event={"ID":"3184235b-9379-4b54-803d-dbb52582e06b","Type":"ContainerDied","Data":"c4feb8e59691a866aa029f4ac85d391f25603c66aeefe0bae491495d70f46b87"} Oct 02 10:00:17 crc kubenswrapper[4771]: I1002 10:00:17.035777 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 02 10:00:17 crc kubenswrapper[4771]: E1002 10:00:17.443774 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb11e684_5673_479e_873b_d61930fe2d52.slice/crio-739ab865b327a7cea11266222255ee0029fb2d9797cdf176b27e6f0ff2e3a925.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.090050 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"109572b4-1dd0-4e0e-a02f-f307c07ec065","Type":"ContainerStarted","Data":"80cc5b3e2368d16a7f37f87c4e26f2f5a136948cd084700f53ba5e3d7f5b96cd"} Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.101826 4771 generic.go:334] "Generic (PLEG): container finished" podID="eb11e684-5673-479e-873b-d61930fe2d52" containerID="739ab865b327a7cea11266222255ee0029fb2d9797cdf176b27e6f0ff2e3a925" exitCode=0 Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.101894 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" event={"ID":"eb11e684-5673-479e-873b-d61930fe2d52","Type":"ContainerDied","Data":"739ab865b327a7cea11266222255ee0029fb2d9797cdf176b27e6f0ff2e3a925"} Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.117748 4771 generic.go:334] "Generic (PLEG): container finished" podID="851bcee9-3b3a-46e9-a950-fcf6cee435ba" containerID="35409fa7746dd7c82183b971b2c77130a66c732a98d5855b35c20ba450df4b8d" exitCode=0 Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.117849 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" event={"ID":"851bcee9-3b3a-46e9-a950-fcf6cee435ba","Type":"ContainerDied","Data":"35409fa7746dd7c82183b971b2c77130a66c732a98d5855b35c20ba450df4b8d"} Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.136939 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-f725t" event={"ID":"975630d7-f63d-43eb-99c4-a245f4f20c24","Type":"ContainerStarted","Data":"0bf3f374add61e91e09dea2502ed16444da7aec47589d12eb1132079e99b1dec"} Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.210184 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-f725t" podStartSLOduration=6.210159934 podStartE2EDuration="6.210159934s" podCreationTimestamp="2025-10-02 10:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:18.168902929 +0000 UTC m=+1405.816588016" watchObservedRunningTime="2025-10-02 10:00:18.210159934 +0000 UTC m=+1405.857845001" Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.229771 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-x9s6s" podUID="5dcb591a-9c0b-44ce-a0f0-db26883d5479" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.177:5353: i/o timeout" Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.623160 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.720991 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-dns-swift-storage-0\") pod \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.721449 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-config\") pod \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.721523 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-ovsdbserver-sb\") pod \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.721584 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcng4\" (UniqueName: \"kubernetes.io/projected/851bcee9-3b3a-46e9-a950-fcf6cee435ba-kube-api-access-zcng4\") pod \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.721679 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-dns-svc\") pod \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.721764 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-ovsdbserver-nb\") pod \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\" (UID: \"851bcee9-3b3a-46e9-a950-fcf6cee435ba\") " Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.741497 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/851bcee9-3b3a-46e9-a950-fcf6cee435ba-kube-api-access-zcng4" (OuterVolumeSpecName: "kube-api-access-zcng4") pod "851bcee9-3b3a-46e9-a950-fcf6cee435ba" (UID: "851bcee9-3b3a-46e9-a950-fcf6cee435ba"). InnerVolumeSpecName "kube-api-access-zcng4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:18 crc kubenswrapper[4771]: I1002 10:00:18.828838 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcng4\" (UniqueName: \"kubernetes.io/projected/851bcee9-3b3a-46e9-a950-fcf6cee435ba-kube-api-access-zcng4\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.079346 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-config" (OuterVolumeSpecName: "config") pod "851bcee9-3b3a-46e9-a950-fcf6cee435ba" (UID: "851bcee9-3b3a-46e9-a950-fcf6cee435ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.196336 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.251244 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "851bcee9-3b3a-46e9-a950-fcf6cee435ba" (UID: "851bcee9-3b3a-46e9-a950-fcf6cee435ba"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.268397 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "851bcee9-3b3a-46e9-a950-fcf6cee435ba" (UID: "851bcee9-3b3a-46e9-a950-fcf6cee435ba"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.279962 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" event={"ID":"851bcee9-3b3a-46e9-a950-fcf6cee435ba","Type":"ContainerDied","Data":"62b2b48f464692140961535ec4384440f6747b6fc2a284b44f0e498b554dd91a"} Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.280048 4771 scope.go:117] "RemoveContainer" containerID="35409fa7746dd7c82183b971b2c77130a66c732a98d5855b35c20ba450df4b8d" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.280228 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-vdt6g" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.300258 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.300302 4771 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.319311 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4c5j" event={"ID":"3184235b-9379-4b54-803d-dbb52582e06b","Type":"ContainerStarted","Data":"9dd20d380201ee099ec8bcaf1a7d049eb27504b670bd8087bb70301b7a4509b0"} Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.357362 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "851bcee9-3b3a-46e9-a950-fcf6cee435ba" (UID: "851bcee9-3b3a-46e9-a950-fcf6cee435ba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.364619 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "851bcee9-3b3a-46e9-a950-fcf6cee435ba" (UID: "851bcee9-3b3a-46e9-a950-fcf6cee435ba"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.368054 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fea444ce-0980-4dd4-a4d4-f391e79f6a05","Type":"ContainerStarted","Data":"99d35e80fd34d432e6dd8b742959b357ac767634dcb72b0ca8321a47b7746f57"} Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.387401 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j4c5j" podStartSLOduration=3.5716100109999998 podStartE2EDuration="10.387371307s" podCreationTimestamp="2025-10-02 10:00:09 +0000 UTC" firstStartedPulling="2025-10-02 10:00:11.40658181 +0000 UTC m=+1399.054266877" lastFinishedPulling="2025-10-02 10:00:18.222343106 +0000 UTC m=+1405.870028173" observedRunningTime="2025-10-02 10:00:19.365400395 +0000 UTC m=+1407.013085462" watchObservedRunningTime="2025-10-02 10:00:19.387371307 +0000 UTC m=+1407.035056374" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.403011 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.403060 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/851bcee9-3b3a-46e9-a950-fcf6cee435ba-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.744209 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-vdt6g"] Oct 02 10:00:19 crc kubenswrapper[4771]: I1002 10:00:19.787943 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-vdt6g"] Oct 02 10:00:20 crc kubenswrapper[4771]: I1002 10:00:20.027266 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:00:20 crc kubenswrapper[4771]: I1002 10:00:20.027802 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:00:20 crc kubenswrapper[4771]: I1002 10:00:20.421946 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"109572b4-1dd0-4e0e-a02f-f307c07ec065","Type":"ContainerStarted","Data":"e2ab5b67bdc92e37642013d2774beea45717953458f50b2909955d46f27bf92a"} Oct 02 10:00:20 crc kubenswrapper[4771]: I1002 10:00:20.422702 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="109572b4-1dd0-4e0e-a02f-f307c07ec065" containerName="glance-log" containerID="cri-o://80cc5b3e2368d16a7f37f87c4e26f2f5a136948cd084700f53ba5e3d7f5b96cd" gracePeriod=30 Oct 02 10:00:20 crc kubenswrapper[4771]: I1002 10:00:20.423639 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="109572b4-1dd0-4e0e-a02f-f307c07ec065" containerName="glance-httpd" containerID="cri-o://e2ab5b67bdc92e37642013d2774beea45717953458f50b2909955d46f27bf92a" gracePeriod=30 Oct 02 10:00:20 crc kubenswrapper[4771]: I1002 10:00:20.435694 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" event={"ID":"eb11e684-5673-479e-873b-d61930fe2d52","Type":"ContainerStarted","Data":"31ea447f7214766a6d3dd9b45518a5143478760e9e91f9ee76e181ab8b6af100"} Oct 02 10:00:20 crc kubenswrapper[4771]: I1002 10:00:20.437508 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:20 crc kubenswrapper[4771]: I1002 10:00:20.466570 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.466542152 podStartE2EDuration="8.466542152s" podCreationTimestamp="2025-10-02 10:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:20.444864528 +0000 UTC m=+1408.092549605" watchObservedRunningTime="2025-10-02 10:00:20.466542152 +0000 UTC m=+1408.114227219" Oct 02 10:00:20 crc kubenswrapper[4771]: I1002 10:00:20.498562 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" podStartSLOduration=8.49853645 podStartE2EDuration="8.49853645s" podCreationTimestamp="2025-10-02 10:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:20.489718455 +0000 UTC m=+1408.137403522" watchObservedRunningTime="2025-10-02 10:00:20.49853645 +0000 UTC m=+1408.146221517" Oct 02 10:00:21 crc kubenswrapper[4771]: I1002 10:00:21.202152 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j4c5j" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="registry-server" probeResult="failure" output=< Oct 02 10:00:21 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:00:21 crc kubenswrapper[4771]: > Oct 02 10:00:21 crc kubenswrapper[4771]: I1002 10:00:21.517727 4771 generic.go:334] "Generic (PLEG): container finished" podID="109572b4-1dd0-4e0e-a02f-f307c07ec065" containerID="e2ab5b67bdc92e37642013d2774beea45717953458f50b2909955d46f27bf92a" exitCode=143 Oct 02 10:00:21 crc kubenswrapper[4771]: I1002 10:00:21.517763 4771 generic.go:334] "Generic (PLEG): container finished" podID="109572b4-1dd0-4e0e-a02f-f307c07ec065" containerID="80cc5b3e2368d16a7f37f87c4e26f2f5a136948cd084700f53ba5e3d7f5b96cd" exitCode=143 Oct 02 10:00:21 crc kubenswrapper[4771]: I1002 10:00:21.517801 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"109572b4-1dd0-4e0e-a02f-f307c07ec065","Type":"ContainerDied","Data":"e2ab5b67bdc92e37642013d2774beea45717953458f50b2909955d46f27bf92a"} Oct 02 10:00:21 crc kubenswrapper[4771]: I1002 10:00:21.517881 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"109572b4-1dd0-4e0e-a02f-f307c07ec065","Type":"ContainerDied","Data":"80cc5b3e2368d16a7f37f87c4e26f2f5a136948cd084700f53ba5e3d7f5b96cd"} Oct 02 10:00:21 crc kubenswrapper[4771]: I1002 10:00:21.521424 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fea444ce-0980-4dd4-a4d4-f391e79f6a05","Type":"ContainerStarted","Data":"a8836e0ff391bbb576c39c46c4a72eccf112ea88d52a9fe3860b72c4c13ca6f9"} Oct 02 10:00:21 crc kubenswrapper[4771]: I1002 10:00:21.728446 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="851bcee9-3b3a-46e9-a950-fcf6cee435ba" path="/var/lib/kubelet/pods/851bcee9-3b3a-46e9-a950-fcf6cee435ba/volumes" Oct 02 10:00:22 crc kubenswrapper[4771]: I1002 10:00:22.575187 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="fea444ce-0980-4dd4-a4d4-f391e79f6a05" containerName="glance-log" containerID="cri-o://99d35e80fd34d432e6dd8b742959b357ac767634dcb72b0ca8321a47b7746f57" gracePeriod=30 Oct 02 10:00:22 crc kubenswrapper[4771]: I1002 10:00:22.580296 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="fea444ce-0980-4dd4-a4d4-f391e79f6a05" containerName="glance-httpd" containerID="cri-o://a8836e0ff391bbb576c39c46c4a72eccf112ea88d52a9fe3860b72c4c13ca6f9" gracePeriod=30 Oct 02 10:00:22 crc kubenswrapper[4771]: I1002 10:00:22.738761 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.738719454 podStartE2EDuration="11.738719454s" podCreationTimestamp="2025-10-02 10:00:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:22.734586708 +0000 UTC m=+1410.382271775" watchObservedRunningTime="2025-10-02 10:00:22.738719454 +0000 UTC m=+1410.386404531" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.216079 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.261597 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"109572b4-1dd0-4e0e-a02f-f307c07ec065\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.261684 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-combined-ca-bundle\") pod \"109572b4-1dd0-4e0e-a02f-f307c07ec065\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.261763 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-config-data\") pod \"109572b4-1dd0-4e0e-a02f-f307c07ec065\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.261808 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/109572b4-1dd0-4e0e-a02f-f307c07ec065-logs\") pod \"109572b4-1dd0-4e0e-a02f-f307c07ec065\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.261867 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgmhn\" (UniqueName: \"kubernetes.io/projected/109572b4-1dd0-4e0e-a02f-f307c07ec065-kube-api-access-dgmhn\") pod \"109572b4-1dd0-4e0e-a02f-f307c07ec065\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.261914 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-scripts\") pod \"109572b4-1dd0-4e0e-a02f-f307c07ec065\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.261936 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/109572b4-1dd0-4e0e-a02f-f307c07ec065-httpd-run\") pod \"109572b4-1dd0-4e0e-a02f-f307c07ec065\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.265043 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/109572b4-1dd0-4e0e-a02f-f307c07ec065-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "109572b4-1dd0-4e0e-a02f-f307c07ec065" (UID: "109572b4-1dd0-4e0e-a02f-f307c07ec065"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.266022 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/109572b4-1dd0-4e0e-a02f-f307c07ec065-logs" (OuterVolumeSpecName: "logs") pod "109572b4-1dd0-4e0e-a02f-f307c07ec065" (UID: "109572b4-1dd0-4e0e-a02f-f307c07ec065"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.271735 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "109572b4-1dd0-4e0e-a02f-f307c07ec065" (UID: "109572b4-1dd0-4e0e-a02f-f307c07ec065"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.274116 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-scripts" (OuterVolumeSpecName: "scripts") pod "109572b4-1dd0-4e0e-a02f-f307c07ec065" (UID: "109572b4-1dd0-4e0e-a02f-f307c07ec065"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.276264 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/109572b4-1dd0-4e0e-a02f-f307c07ec065-kube-api-access-dgmhn" (OuterVolumeSpecName: "kube-api-access-dgmhn") pod "109572b4-1dd0-4e0e-a02f-f307c07ec065" (UID: "109572b4-1dd0-4e0e-a02f-f307c07ec065"). InnerVolumeSpecName "kube-api-access-dgmhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.312259 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "109572b4-1dd0-4e0e-a02f-f307c07ec065" (UID: "109572b4-1dd0-4e0e-a02f-f307c07ec065"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.365327 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-internal-tls-certs\") pod \"109572b4-1dd0-4e0e-a02f-f307c07ec065\" (UID: \"109572b4-1dd0-4e0e-a02f-f307c07ec065\") " Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.365691 4771 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/109572b4-1dd0-4e0e-a02f-f307c07ec065-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.365704 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgmhn\" (UniqueName: \"kubernetes.io/projected/109572b4-1dd0-4e0e-a02f-f307c07ec065-kube-api-access-dgmhn\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.365715 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.365725 4771 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/109572b4-1dd0-4e0e-a02f-f307c07ec065-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.365747 4771 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.365756 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.382423 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-config-data" (OuterVolumeSpecName: "config-data") pod "109572b4-1dd0-4e0e-a02f-f307c07ec065" (UID: "109572b4-1dd0-4e0e-a02f-f307c07ec065"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.392357 4771 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.443044 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "109572b4-1dd0-4e0e-a02f-f307c07ec065" (UID: "109572b4-1dd0-4e0e-a02f-f307c07ec065"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.467857 4771 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.467912 4771 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.467933 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109572b4-1dd0-4e0e-a02f-f307c07ec065-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.594851 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"109572b4-1dd0-4e0e-a02f-f307c07ec065","Type":"ContainerDied","Data":"e4659721bfcda964595def3f51e014567bf2e9b2a437a03b749c4a25c8992d63"} Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.594902 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.594913 4771 scope.go:117] "RemoveContainer" containerID="e2ab5b67bdc92e37642013d2774beea45717953458f50b2909955d46f27bf92a" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.600987 4771 generic.go:334] "Generic (PLEG): container finished" podID="ec499dd8-fc17-46bd-8771-eb78b728da89" containerID="64a69a71b8ee1037e5918d2efd2826cac3d09f55d8d7bc0c97a2614cb0f260b3" exitCode=0 Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.601104 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bm9ln" event={"ID":"ec499dd8-fc17-46bd-8771-eb78b728da89","Type":"ContainerDied","Data":"64a69a71b8ee1037e5918d2efd2826cac3d09f55d8d7bc0c97a2614cb0f260b3"} Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.606234 4771 generic.go:334] "Generic (PLEG): container finished" podID="fea444ce-0980-4dd4-a4d4-f391e79f6a05" containerID="a8836e0ff391bbb576c39c46c4a72eccf112ea88d52a9fe3860b72c4c13ca6f9" exitCode=0 Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.606274 4771 generic.go:334] "Generic (PLEG): container finished" podID="fea444ce-0980-4dd4-a4d4-f391e79f6a05" containerID="99d35e80fd34d432e6dd8b742959b357ac767634dcb72b0ca8321a47b7746f57" exitCode=143 Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.606273 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fea444ce-0980-4dd4-a4d4-f391e79f6a05","Type":"ContainerDied","Data":"a8836e0ff391bbb576c39c46c4a72eccf112ea88d52a9fe3860b72c4c13ca6f9"} Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.606339 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fea444ce-0980-4dd4-a4d4-f391e79f6a05","Type":"ContainerDied","Data":"99d35e80fd34d432e6dd8b742959b357ac767634dcb72b0ca8321a47b7746f57"} Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.673267 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.726893 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.726951 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:00:23 crc kubenswrapper[4771]: E1002 10:00:23.727380 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="851bcee9-3b3a-46e9-a950-fcf6cee435ba" containerName="init" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.727396 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="851bcee9-3b3a-46e9-a950-fcf6cee435ba" containerName="init" Oct 02 10:00:23 crc kubenswrapper[4771]: E1002 10:00:23.727413 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="109572b4-1dd0-4e0e-a02f-f307c07ec065" containerName="glance-log" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.727419 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="109572b4-1dd0-4e0e-a02f-f307c07ec065" containerName="glance-log" Oct 02 10:00:23 crc kubenswrapper[4771]: E1002 10:00:23.727449 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="109572b4-1dd0-4e0e-a02f-f307c07ec065" containerName="glance-httpd" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.727454 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="109572b4-1dd0-4e0e-a02f-f307c07ec065" containerName="glance-httpd" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.727675 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="851bcee9-3b3a-46e9-a950-fcf6cee435ba" containerName="init" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.727694 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="109572b4-1dd0-4e0e-a02f-f307c07ec065" containerName="glance-httpd" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.727712 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="109572b4-1dd0-4e0e-a02f-f307c07ec065" containerName="glance-log" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.731267 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.734336 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.736521 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.775155 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.881617 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.881960 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.882100 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.882213 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ecc1a24-b566-4f4b-8a39-10b49533cd66-logs\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.882349 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.882439 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4ecc1a24-b566-4f4b-8a39-10b49533cd66-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.882556 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.882662 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzncw\" (UniqueName: \"kubernetes.io/projected/4ecc1a24-b566-4f4b-8a39-10b49533cd66-kube-api-access-qzncw\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.985610 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.986080 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.986168 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.986194 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ecc1a24-b566-4f4b-8a39-10b49533cd66-logs\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.986247 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.986281 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4ecc1a24-b566-4f4b-8a39-10b49533cd66-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.986317 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.986364 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzncw\" (UniqueName: \"kubernetes.io/projected/4ecc1a24-b566-4f4b-8a39-10b49533cd66-kube-api-access-qzncw\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.986812 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ecc1a24-b566-4f4b-8a39-10b49533cd66-logs\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.986938 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.987342 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4ecc1a24-b566-4f4b-8a39-10b49533cd66-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.992030 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.992397 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:23 crc kubenswrapper[4771]: I1002 10:00:23.994582 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:24 crc kubenswrapper[4771]: I1002 10:00:24.000511 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:24 crc kubenswrapper[4771]: I1002 10:00:24.010091 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzncw\" (UniqueName: \"kubernetes.io/projected/4ecc1a24-b566-4f4b-8a39-10b49533cd66-kube-api-access-qzncw\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:24 crc kubenswrapper[4771]: I1002 10:00:24.027510 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:00:24 crc kubenswrapper[4771]: I1002 10:00:24.070293 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:25 crc kubenswrapper[4771]: I1002 10:00:25.698120 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="109572b4-1dd0-4e0e-a02f-f307c07ec065" path="/var/lib/kubelet/pods/109572b4-1dd0-4e0e-a02f-f307c07ec065/volumes" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.017693 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.142791 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-combined-ca-bundle\") pod \"ec499dd8-fc17-46bd-8771-eb78b728da89\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.142991 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-config-data\") pod \"ec499dd8-fc17-46bd-8771-eb78b728da89\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.143065 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-credential-keys\") pod \"ec499dd8-fc17-46bd-8771-eb78b728da89\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.143120 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58qnb\" (UniqueName: \"kubernetes.io/projected/ec499dd8-fc17-46bd-8771-eb78b728da89-kube-api-access-58qnb\") pod \"ec499dd8-fc17-46bd-8771-eb78b728da89\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.143215 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-fernet-keys\") pod \"ec499dd8-fc17-46bd-8771-eb78b728da89\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.143297 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-scripts\") pod \"ec499dd8-fc17-46bd-8771-eb78b728da89\" (UID: \"ec499dd8-fc17-46bd-8771-eb78b728da89\") " Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.153015 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ec499dd8-fc17-46bd-8771-eb78b728da89" (UID: "ec499dd8-fc17-46bd-8771-eb78b728da89"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.153820 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-scripts" (OuterVolumeSpecName: "scripts") pod "ec499dd8-fc17-46bd-8771-eb78b728da89" (UID: "ec499dd8-fc17-46bd-8771-eb78b728da89"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.155865 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec499dd8-fc17-46bd-8771-eb78b728da89-kube-api-access-58qnb" (OuterVolumeSpecName: "kube-api-access-58qnb") pod "ec499dd8-fc17-46bd-8771-eb78b728da89" (UID: "ec499dd8-fc17-46bd-8771-eb78b728da89"). InnerVolumeSpecName "kube-api-access-58qnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.157471 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ec499dd8-fc17-46bd-8771-eb78b728da89" (UID: "ec499dd8-fc17-46bd-8771-eb78b728da89"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.183243 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-config-data" (OuterVolumeSpecName: "config-data") pod "ec499dd8-fc17-46bd-8771-eb78b728da89" (UID: "ec499dd8-fc17-46bd-8771-eb78b728da89"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.192155 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec499dd8-fc17-46bd-8771-eb78b728da89" (UID: "ec499dd8-fc17-46bd-8771-eb78b728da89"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.246979 4771 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.247033 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.247045 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.247061 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.247073 4771 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ec499dd8-fc17-46bd-8771-eb78b728da89-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.247087 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58qnb\" (UniqueName: \"kubernetes.io/projected/ec499dd8-fc17-46bd-8771-eb78b728da89-kube-api-access-58qnb\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.681658 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bm9ln" event={"ID":"ec499dd8-fc17-46bd-8771-eb78b728da89","Type":"ContainerDied","Data":"543aeeb78a6ffd39a616fe47a54200a595f2ef6fcbbfaebbb7808d5f7026582d"} Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.681734 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="543aeeb78a6ffd39a616fe47a54200a595f2ef6fcbbfaebbb7808d5f7026582d" Oct 02 10:00:26 crc kubenswrapper[4771]: I1002 10:00:26.681756 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bm9ln" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.115044 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-bm9ln"] Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.127711 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-bm9ln"] Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.209849 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-6gz2g"] Oct 02 10:00:27 crc kubenswrapper[4771]: E1002 10:00:27.210466 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec499dd8-fc17-46bd-8771-eb78b728da89" containerName="keystone-bootstrap" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.210503 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec499dd8-fc17-46bd-8771-eb78b728da89" containerName="keystone-bootstrap" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.210848 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec499dd8-fc17-46bd-8771-eb78b728da89" containerName="keystone-bootstrap" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.212662 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.215600 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.216032 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.216100 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8kg7s" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.216041 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.227663 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6gz2g"] Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.391334 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-combined-ca-bundle\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.391407 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvn8w\" (UniqueName: \"kubernetes.io/projected/068a6ab1-1a57-489b-adf2-67cb4011216e-kube-api-access-xvn8w\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.391443 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-credential-keys\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.391515 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-scripts\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.391546 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-fernet-keys\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.391609 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-config-data\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.502036 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-config-data\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.502181 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-combined-ca-bundle\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.502230 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvn8w\" (UniqueName: \"kubernetes.io/projected/068a6ab1-1a57-489b-adf2-67cb4011216e-kube-api-access-xvn8w\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.502260 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-credential-keys\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.502345 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-scripts\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.502418 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-fernet-keys\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.509982 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-config-data\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.512420 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-fernet-keys\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.513492 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-combined-ca-bundle\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.515941 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-credential-keys\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.524497 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-scripts\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.524826 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvn8w\" (UniqueName: \"kubernetes.io/projected/068a6ab1-1a57-489b-adf2-67cb4011216e-kube-api-access-xvn8w\") pod \"keystone-bootstrap-6gz2g\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.543366 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:27 crc kubenswrapper[4771]: I1002 10:00:27.704686 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec499dd8-fc17-46bd-8771-eb78b728da89" path="/var/lib/kubelet/pods/ec499dd8-fc17-46bd-8771-eb78b728da89/volumes" Oct 02 10:00:28 crc kubenswrapper[4771]: I1002 10:00:28.814024 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:00:28 crc kubenswrapper[4771]: I1002 10:00:28.898428 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-bjg2r"] Oct 02 10:00:28 crc kubenswrapper[4771]: I1002 10:00:28.899030 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-bjg2r" podUID="e786364e-fd43-43ed-a371-b87cc5a1e59a" containerName="dnsmasq-dns" containerID="cri-o://db107d781bec174e1922a04a1d846e1cb676be4c68153f675441a3c53dc84bdb" gracePeriod=10 Oct 02 10:00:29 crc kubenswrapper[4771]: I1002 10:00:29.102529 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-bjg2r" podUID="e786364e-fd43-43ed-a371-b87cc5a1e59a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.153:5353: connect: connection refused" Oct 02 10:00:29 crc kubenswrapper[4771]: I1002 10:00:29.744513 4771 generic.go:334] "Generic (PLEG): container finished" podID="e786364e-fd43-43ed-a371-b87cc5a1e59a" containerID="db107d781bec174e1922a04a1d846e1cb676be4c68153f675441a3c53dc84bdb" exitCode=0 Oct 02 10:00:29 crc kubenswrapper[4771]: I1002 10:00:29.744564 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-bjg2r" event={"ID":"e786364e-fd43-43ed-a371-b87cc5a1e59a","Type":"ContainerDied","Data":"db107d781bec174e1922a04a1d846e1cb676be4c68153f675441a3c53dc84bdb"} Oct 02 10:00:31 crc kubenswrapper[4771]: I1002 10:00:31.073667 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j4c5j" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="registry-server" probeResult="failure" output=< Oct 02 10:00:31 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:00:31 crc kubenswrapper[4771]: > Oct 02 10:00:32 crc kubenswrapper[4771]: I1002 10:00:32.410172 4771 scope.go:117] "RemoveContainer" containerID="80cc5b3e2368d16a7f37f87c4e26f2f5a136948cd084700f53ba5e3d7f5b96cd" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.102674 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-bjg2r" podUID="e786364e-fd43-43ed-a371-b87cc5a1e59a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.153:5353: connect: connection refused" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.352360 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.515619 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.515803 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6bxz\" (UniqueName: \"kubernetes.io/projected/fea444ce-0980-4dd4-a4d4-f391e79f6a05-kube-api-access-r6bxz\") pod \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.515925 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fea444ce-0980-4dd4-a4d4-f391e79f6a05-httpd-run\") pod \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.515971 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-public-tls-certs\") pod \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.516002 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-config-data\") pod \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.516203 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-combined-ca-bundle\") pod \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.516315 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fea444ce-0980-4dd4-a4d4-f391e79f6a05-logs\") pod \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.516399 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fea444ce-0980-4dd4-a4d4-f391e79f6a05-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fea444ce-0980-4dd4-a4d4-f391e79f6a05" (UID: "fea444ce-0980-4dd4-a4d4-f391e79f6a05"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.516635 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-scripts\") pod \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\" (UID: \"fea444ce-0980-4dd4-a4d4-f391e79f6a05\") " Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.516744 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fea444ce-0980-4dd4-a4d4-f391e79f6a05-logs" (OuterVolumeSpecName: "logs") pod "fea444ce-0980-4dd4-a4d4-f391e79f6a05" (UID: "fea444ce-0980-4dd4-a4d4-f391e79f6a05"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.517226 4771 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fea444ce-0980-4dd4-a4d4-f391e79f6a05-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.517247 4771 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fea444ce-0980-4dd4-a4d4-f391e79f6a05-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.539380 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-scripts" (OuterVolumeSpecName: "scripts") pod "fea444ce-0980-4dd4-a4d4-f391e79f6a05" (UID: "fea444ce-0980-4dd4-a4d4-f391e79f6a05"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.539439 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fea444ce-0980-4dd4-a4d4-f391e79f6a05-kube-api-access-r6bxz" (OuterVolumeSpecName: "kube-api-access-r6bxz") pod "fea444ce-0980-4dd4-a4d4-f391e79f6a05" (UID: "fea444ce-0980-4dd4-a4d4-f391e79f6a05"). InnerVolumeSpecName "kube-api-access-r6bxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.548110 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "fea444ce-0980-4dd4-a4d4-f391e79f6a05" (UID: "fea444ce-0980-4dd4-a4d4-f391e79f6a05"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.555876 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fea444ce-0980-4dd4-a4d4-f391e79f6a05" (UID: "fea444ce-0980-4dd4-a4d4-f391e79f6a05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.584227 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-config-data" (OuterVolumeSpecName: "config-data") pod "fea444ce-0980-4dd4-a4d4-f391e79f6a05" (UID: "fea444ce-0980-4dd4-a4d4-f391e79f6a05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.602901 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "fea444ce-0980-4dd4-a4d4-f391e79f6a05" (UID: "fea444ce-0980-4dd4-a4d4-f391e79f6a05"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.619701 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.619772 4771 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.619790 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6bxz\" (UniqueName: \"kubernetes.io/projected/fea444ce-0980-4dd4-a4d4-f391e79f6a05-kube-api-access-r6bxz\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.619809 4771 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.619825 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.619838 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fea444ce-0980-4dd4-a4d4-f391e79f6a05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.653791 4771 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.723772 4771 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.810165 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fea444ce-0980-4dd4-a4d4-f391e79f6a05","Type":"ContainerDied","Data":"a4fb157d30ed972b147d9bd1698fffce6c06106f3b96ee843636df0eb88e09e0"} Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.810413 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.856708 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.890266 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.899259 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:00:34 crc kubenswrapper[4771]: E1002 10:00:34.899924 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fea444ce-0980-4dd4-a4d4-f391e79f6a05" containerName="glance-log" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.899943 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="fea444ce-0980-4dd4-a4d4-f391e79f6a05" containerName="glance-log" Oct 02 10:00:34 crc kubenswrapper[4771]: E1002 10:00:34.899985 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fea444ce-0980-4dd4-a4d4-f391e79f6a05" containerName="glance-httpd" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.899992 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="fea444ce-0980-4dd4-a4d4-f391e79f6a05" containerName="glance-httpd" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.900254 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="fea444ce-0980-4dd4-a4d4-f391e79f6a05" containerName="glance-httpd" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.900289 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="fea444ce-0980-4dd4-a4d4-f391e79f6a05" containerName="glance-log" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.901846 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.905456 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.905639 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 10:00:34 crc kubenswrapper[4771]: I1002 10:00:34.938365 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.038572 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69624004-f6bf-48f2-af01-44b450260c6a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.038735 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-config-data\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.038781 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.038812 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.038945 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-scripts\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.039030 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.039206 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh4hc\" (UniqueName: \"kubernetes.io/projected/69624004-f6bf-48f2-af01-44b450260c6a-kube-api-access-jh4hc\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.039507 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69624004-f6bf-48f2-af01-44b450260c6a-logs\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.142997 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh4hc\" (UniqueName: \"kubernetes.io/projected/69624004-f6bf-48f2-af01-44b450260c6a-kube-api-access-jh4hc\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.143072 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69624004-f6bf-48f2-af01-44b450260c6a-logs\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.143162 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69624004-f6bf-48f2-af01-44b450260c6a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.143204 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-config-data\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.143235 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.143265 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.143334 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-scripts\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.143402 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.146060 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.146695 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69624004-f6bf-48f2-af01-44b450260c6a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.147077 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69624004-f6bf-48f2-af01-44b450260c6a-logs\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.148892 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.149347 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-scripts\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.150393 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-config-data\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.158172 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.164160 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh4hc\" (UniqueName: \"kubernetes.io/projected/69624004-f6bf-48f2-af01-44b450260c6a-kube-api-access-jh4hc\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.195238 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.235968 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:00:35 crc kubenswrapper[4771]: I1002 10:00:35.696804 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fea444ce-0980-4dd4-a4d4-f391e79f6a05" path="/var/lib/kubelet/pods/fea444ce-0980-4dd4-a4d4-f391e79f6a05/volumes" Oct 02 10:00:39 crc kubenswrapper[4771]: I1002 10:00:39.101918 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-bjg2r" podUID="e786364e-fd43-43ed-a371-b87cc5a1e59a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.153:5353: connect: connection refused" Oct 02 10:00:39 crc kubenswrapper[4771]: I1002 10:00:39.102098 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 10:00:41 crc kubenswrapper[4771]: I1002 10:00:41.076390 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j4c5j" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="registry-server" probeResult="failure" output=< Oct 02 10:00:41 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:00:41 crc kubenswrapper[4771]: > Oct 02 10:00:44 crc kubenswrapper[4771]: I1002 10:00:44.102053 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-bjg2r" podUID="e786364e-fd43-43ed-a371-b87cc5a1e59a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.153:5353: connect: connection refused" Oct 02 10:00:44 crc kubenswrapper[4771]: E1002 10:00:44.994667 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 02 10:00:44 crc kubenswrapper[4771]: E1002 10:00:44.995217 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-khdp2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-wg2x2_openstack(a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:00:44 crc kubenswrapper[4771]: E1002 10:00:44.996893 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-wg2x2" podUID="a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5" Oct 02 10:00:45 crc kubenswrapper[4771]: E1002 10:00:45.741080 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 02 10:00:45 crc kubenswrapper[4771]: E1002 10:00:45.741545 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9wv5l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-5dxhb_openstack(3bc64084-015e-47d2-b20c-f241e367fe27): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:00:45 crc kubenswrapper[4771]: E1002 10:00:45.743264 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-5dxhb" podUID="3bc64084-015e-47d2-b20c-f241e367fe27" Oct 02 10:00:45 crc kubenswrapper[4771]: E1002 10:00:45.966918 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-wg2x2" podUID="a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5" Oct 02 10:00:45 crc kubenswrapper[4771]: E1002 10:00:45.967000 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-5dxhb" podUID="3bc64084-015e-47d2-b20c-f241e367fe27" Oct 02 10:00:46 crc kubenswrapper[4771]: I1002 10:00:46.843074 4771 scope.go:117] "RemoveContainer" containerID="a8836e0ff391bbb576c39c46c4a72eccf112ea88d52a9fe3860b72c4c13ca6f9" Oct 02 10:00:46 crc kubenswrapper[4771]: E1002 10:00:46.889048 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 02 10:00:46 crc kubenswrapper[4771]: E1002 10:00:46.889346 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-twr2n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-pkkwb_openstack(2a276f13-9f48-40ae-8445-ce376fce62a1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:00:46 crc kubenswrapper[4771]: E1002 10:00:46.893505 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-pkkwb" podUID="2a276f13-9f48-40ae-8445-ce376fce62a1" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.057784 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-bjg2r" event={"ID":"e786364e-fd43-43ed-a371-b87cc5a1e59a","Type":"ContainerDied","Data":"c6d8b754cd91ab483faff42b6c009f6abc6b3d35e4821bded38e2112de0153cd"} Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.058313 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6d8b754cd91ab483faff42b6c009f6abc6b3d35e4821bded38e2112de0153cd" Oct 02 10:00:47 crc kubenswrapper[4771]: E1002 10:00:47.069535 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-pkkwb" podUID="2a276f13-9f48-40ae-8445-ce376fce62a1" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.204628 4771 scope.go:117] "RemoveContainer" containerID="99d35e80fd34d432e6dd8b742959b357ac767634dcb72b0ca8321a47b7746f57" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.308807 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.468388 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-ovsdbserver-sb\") pod \"e786364e-fd43-43ed-a371-b87cc5a1e59a\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.468465 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-config\") pod \"e786364e-fd43-43ed-a371-b87cc5a1e59a\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.468734 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2lw8\" (UniqueName: \"kubernetes.io/projected/e786364e-fd43-43ed-a371-b87cc5a1e59a-kube-api-access-j2lw8\") pod \"e786364e-fd43-43ed-a371-b87cc5a1e59a\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.468804 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-dns-svc\") pod \"e786364e-fd43-43ed-a371-b87cc5a1e59a\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.468858 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-ovsdbserver-nb\") pod \"e786364e-fd43-43ed-a371-b87cc5a1e59a\" (UID: \"e786364e-fd43-43ed-a371-b87cc5a1e59a\") " Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.477372 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e786364e-fd43-43ed-a371-b87cc5a1e59a-kube-api-access-j2lw8" (OuterVolumeSpecName: "kube-api-access-j2lw8") pod "e786364e-fd43-43ed-a371-b87cc5a1e59a" (UID: "e786364e-fd43-43ed-a371-b87cc5a1e59a"). InnerVolumeSpecName "kube-api-access-j2lw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.531615 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.551267 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e786364e-fd43-43ed-a371-b87cc5a1e59a" (UID: "e786364e-fd43-43ed-a371-b87cc5a1e59a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.552835 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6gz2g"] Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.553551 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-config" (OuterVolumeSpecName: "config") pod "e786364e-fd43-43ed-a371-b87cc5a1e59a" (UID: "e786364e-fd43-43ed-a371-b87cc5a1e59a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.557088 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e786364e-fd43-43ed-a371-b87cc5a1e59a" (UID: "e786364e-fd43-43ed-a371-b87cc5a1e59a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.560535 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e786364e-fd43-43ed-a371-b87cc5a1e59a" (UID: "e786364e-fd43-43ed-a371-b87cc5a1e59a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.577316 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2lw8\" (UniqueName: \"kubernetes.io/projected/e786364e-fd43-43ed-a371-b87cc5a1e59a-kube-api-access-j2lw8\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.577353 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.577367 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.577378 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.577389 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e786364e-fd43-43ed-a371-b87cc5a1e59a-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:47 crc kubenswrapper[4771]: I1002 10:00:47.787866 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:00:47 crc kubenswrapper[4771]: W1002 10:00:47.791842 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69624004_f6bf_48f2_af01_44b450260c6a.slice/crio-1cd4acf95d3a88f76237bf23f4314dd5ba1251d5f3669f036984d96178833cf8 WatchSource:0}: Error finding container 1cd4acf95d3a88f76237bf23f4314dd5ba1251d5f3669f036984d96178833cf8: Status 404 returned error can't find the container with id 1cd4acf95d3a88f76237bf23f4314dd5ba1251d5f3669f036984d96178833cf8 Oct 02 10:00:48 crc kubenswrapper[4771]: I1002 10:00:48.079854 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"48fe997b-eed7-4191-9f63-ecb40112cf50","Type":"ContainerStarted","Data":"db909e1281bba61267bf0467d67eb148eb35688311d23f6a4c31582d35f73211"} Oct 02 10:00:48 crc kubenswrapper[4771]: I1002 10:00:48.088209 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4ecc1a24-b566-4f4b-8a39-10b49533cd66","Type":"ContainerStarted","Data":"509cc0aa180dbfa5b9fc470179aca0e1ccbbc18dd9bb7593d8ffb28ca2072330"} Oct 02 10:00:48 crc kubenswrapper[4771]: I1002 10:00:48.090824 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-dbght" event={"ID":"042a2fb4-504c-4448-922b-48db5507e964","Type":"ContainerStarted","Data":"e0a3421d25863823e2518b80b0336067bf482596d694069ade7893462313fb3a"} Oct 02 10:00:48 crc kubenswrapper[4771]: I1002 10:00:48.094048 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6gz2g" event={"ID":"068a6ab1-1a57-489b-adf2-67cb4011216e","Type":"ContainerStarted","Data":"f89a4c2ff4e398b811223f96db0542f7438cc196051779df68fe2746de06e645"} Oct 02 10:00:48 crc kubenswrapper[4771]: I1002 10:00:48.094086 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6gz2g" event={"ID":"068a6ab1-1a57-489b-adf2-67cb4011216e","Type":"ContainerStarted","Data":"23ddb9843ac4c886198bc23ec7ad98003024e333a272f5a648f2101a9c267cce"} Oct 02 10:00:48 crc kubenswrapper[4771]: I1002 10:00:48.097709 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"69624004-f6bf-48f2-af01-44b450260c6a","Type":"ContainerStarted","Data":"1cd4acf95d3a88f76237bf23f4314dd5ba1251d5f3669f036984d96178833cf8"} Oct 02 10:00:48 crc kubenswrapper[4771]: I1002 10:00:48.099680 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-bjg2r" Oct 02 10:00:48 crc kubenswrapper[4771]: I1002 10:00:48.112279 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-dbght" podStartSLOduration=4.250082179 podStartE2EDuration="36.112261761s" podCreationTimestamp="2025-10-02 10:00:12 +0000 UTC" firstStartedPulling="2025-10-02 10:00:15.002436849 +0000 UTC m=+1402.650121916" lastFinishedPulling="2025-10-02 10:00:46.864616431 +0000 UTC m=+1434.512301498" observedRunningTime="2025-10-02 10:00:48.107093801 +0000 UTC m=+1435.754778868" watchObservedRunningTime="2025-10-02 10:00:48.112261761 +0000 UTC m=+1435.759946828" Oct 02 10:00:48 crc kubenswrapper[4771]: I1002 10:00:48.150743 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-6gz2g" podStartSLOduration=21.150720217 podStartE2EDuration="21.150720217s" podCreationTimestamp="2025-10-02 10:00:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:48.128327054 +0000 UTC m=+1435.776012121" watchObservedRunningTime="2025-10-02 10:00:48.150720217 +0000 UTC m=+1435.798405274" Oct 02 10:00:48 crc kubenswrapper[4771]: I1002 10:00:48.167143 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-bjg2r"] Oct 02 10:00:48 crc kubenswrapper[4771]: I1002 10:00:48.180381 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-bjg2r"] Oct 02 10:00:49 crc kubenswrapper[4771]: I1002 10:00:49.124337 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4ecc1a24-b566-4f4b-8a39-10b49533cd66","Type":"ContainerStarted","Data":"25ace71c32266c58afbd34ba55f7b5a6a6382d12779760e38f072b19f23cb919"} Oct 02 10:00:49 crc kubenswrapper[4771]: I1002 10:00:49.129587 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"69624004-f6bf-48f2-af01-44b450260c6a","Type":"ContainerStarted","Data":"5d9dabe27276c37383ba954078c45957a44316629dc09a6b9c7f3b7d8cb06fa8"} Oct 02 10:00:49 crc kubenswrapper[4771]: I1002 10:00:49.704520 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e786364e-fd43-43ed-a371-b87cc5a1e59a" path="/var/lib/kubelet/pods/e786364e-fd43-43ed-a371-b87cc5a1e59a/volumes" Oct 02 10:00:50 crc kubenswrapper[4771]: I1002 10:00:50.147289 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"48fe997b-eed7-4191-9f63-ecb40112cf50","Type":"ContainerStarted","Data":"32c91bd9bf80e1b4b42d19b0a1d5a478d0fe72de8bdfc148be56fb134cfdcf42"} Oct 02 10:00:50 crc kubenswrapper[4771]: I1002 10:00:50.149423 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4ecc1a24-b566-4f4b-8a39-10b49533cd66","Type":"ContainerStarted","Data":"206ed65503085f9fbdac82b37985708060c35a2884d9991a7324b4d264d3228f"} Oct 02 10:00:50 crc kubenswrapper[4771]: I1002 10:00:50.151604 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"69624004-f6bf-48f2-af01-44b450260c6a","Type":"ContainerStarted","Data":"8a93dc16801ca30b050caf35b39ec4c79462b62b0bc15228517efa1865e06820"} Oct 02 10:00:50 crc kubenswrapper[4771]: I1002 10:00:50.180731 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=27.180703833 podStartE2EDuration="27.180703833s" podCreationTimestamp="2025-10-02 10:00:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:50.171446383 +0000 UTC m=+1437.819131470" watchObservedRunningTime="2025-10-02 10:00:50.180703833 +0000 UTC m=+1437.828388900" Oct 02 10:00:50 crc kubenswrapper[4771]: I1002 10:00:50.220380 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=16.220345821 podStartE2EDuration="16.220345821s" podCreationTimestamp="2025-10-02 10:00:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:50.203837826 +0000 UTC m=+1437.851522893" watchObservedRunningTime="2025-10-02 10:00:50.220345821 +0000 UTC m=+1437.868030888" Oct 02 10:00:51 crc kubenswrapper[4771]: I1002 10:00:51.077341 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j4c5j" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="registry-server" probeResult="failure" output=< Oct 02 10:00:51 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:00:51 crc kubenswrapper[4771]: > Oct 02 10:00:54 crc kubenswrapper[4771]: I1002 10:00:54.072428 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:54 crc kubenswrapper[4771]: I1002 10:00:54.073284 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:54 crc kubenswrapper[4771]: I1002 10:00:54.073309 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:54 crc kubenswrapper[4771]: I1002 10:00:54.073322 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:54 crc kubenswrapper[4771]: I1002 10:00:54.119695 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:54 crc kubenswrapper[4771]: I1002 10:00:54.126788 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:54 crc kubenswrapper[4771]: E1002 10:00:54.534895 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod068a6ab1_1a57_489b_adf2_67cb4011216e.slice/crio-f89a4c2ff4e398b811223f96db0542f7438cc196051779df68fe2746de06e645.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:00:55 crc kubenswrapper[4771]: I1002 10:00:55.227866 4771 generic.go:334] "Generic (PLEG): container finished" podID="068a6ab1-1a57-489b-adf2-67cb4011216e" containerID="f89a4c2ff4e398b811223f96db0542f7438cc196051779df68fe2746de06e645" exitCode=0 Oct 02 10:00:55 crc kubenswrapper[4771]: I1002 10:00:55.228006 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6gz2g" event={"ID":"068a6ab1-1a57-489b-adf2-67cb4011216e","Type":"ContainerDied","Data":"f89a4c2ff4e398b811223f96db0542f7438cc196051779df68fe2746de06e645"} Oct 02 10:00:55 crc kubenswrapper[4771]: I1002 10:00:55.240811 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 10:00:55 crc kubenswrapper[4771]: I1002 10:00:55.241798 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 10:00:55 crc kubenswrapper[4771]: I1002 10:00:55.281779 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 10:00:55 crc kubenswrapper[4771]: I1002 10:00:55.291005 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.244630 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"48fe997b-eed7-4191-9f63-ecb40112cf50","Type":"ContainerStarted","Data":"e9816a93a5fdd5b53097989567d5375baef8e01f9e272c6a9c166cb806660219"} Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.245068 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.245093 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.784156 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.958172 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-combined-ca-bundle\") pod \"068a6ab1-1a57-489b-adf2-67cb4011216e\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.958354 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvn8w\" (UniqueName: \"kubernetes.io/projected/068a6ab1-1a57-489b-adf2-67cb4011216e-kube-api-access-xvn8w\") pod \"068a6ab1-1a57-489b-adf2-67cb4011216e\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.958448 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-fernet-keys\") pod \"068a6ab1-1a57-489b-adf2-67cb4011216e\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.958542 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-credential-keys\") pod \"068a6ab1-1a57-489b-adf2-67cb4011216e\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.958590 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-config-data\") pod \"068a6ab1-1a57-489b-adf2-67cb4011216e\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.958660 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-scripts\") pod \"068a6ab1-1a57-489b-adf2-67cb4011216e\" (UID: \"068a6ab1-1a57-489b-adf2-67cb4011216e\") " Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.967789 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/068a6ab1-1a57-489b-adf2-67cb4011216e-kube-api-access-xvn8w" (OuterVolumeSpecName: "kube-api-access-xvn8w") pod "068a6ab1-1a57-489b-adf2-67cb4011216e" (UID: "068a6ab1-1a57-489b-adf2-67cb4011216e"). InnerVolumeSpecName "kube-api-access-xvn8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.968080 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "068a6ab1-1a57-489b-adf2-67cb4011216e" (UID: "068a6ab1-1a57-489b-adf2-67cb4011216e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.968167 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "068a6ab1-1a57-489b-adf2-67cb4011216e" (UID: "068a6ab1-1a57-489b-adf2-67cb4011216e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:56 crc kubenswrapper[4771]: I1002 10:00:56.969835 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-scripts" (OuterVolumeSpecName: "scripts") pod "068a6ab1-1a57-489b-adf2-67cb4011216e" (UID: "068a6ab1-1a57-489b-adf2-67cb4011216e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.000266 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "068a6ab1-1a57-489b-adf2-67cb4011216e" (UID: "068a6ab1-1a57-489b-adf2-67cb4011216e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.002295 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-config-data" (OuterVolumeSpecName: "config-data") pod "068a6ab1-1a57-489b-adf2-67cb4011216e" (UID: "068a6ab1-1a57-489b-adf2-67cb4011216e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.061557 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvn8w\" (UniqueName: \"kubernetes.io/projected/068a6ab1-1a57-489b-adf2-67cb4011216e-kube-api-access-xvn8w\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.061596 4771 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.061609 4771 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.061617 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.061626 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.061634 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068a6ab1-1a57-489b-adf2-67cb4011216e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.257319 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6gz2g" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.257326 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6gz2g" event={"ID":"068a6ab1-1a57-489b-adf2-67cb4011216e","Type":"ContainerDied","Data":"23ddb9843ac4c886198bc23ec7ad98003024e333a272f5a648f2101a9c267cce"} Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.257404 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23ddb9843ac4c886198bc23ec7ad98003024e333a272f5a648f2101a9c267cce" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.398370 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-758b4b5558-442kz"] Oct 02 10:00:57 crc kubenswrapper[4771]: E1002 10:00:57.399021 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e786364e-fd43-43ed-a371-b87cc5a1e59a" containerName="dnsmasq-dns" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.399042 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e786364e-fd43-43ed-a371-b87cc5a1e59a" containerName="dnsmasq-dns" Oct 02 10:00:57 crc kubenswrapper[4771]: E1002 10:00:57.399056 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e786364e-fd43-43ed-a371-b87cc5a1e59a" containerName="init" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.399066 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e786364e-fd43-43ed-a371-b87cc5a1e59a" containerName="init" Oct 02 10:00:57 crc kubenswrapper[4771]: E1002 10:00:57.399092 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="068a6ab1-1a57-489b-adf2-67cb4011216e" containerName="keystone-bootstrap" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.399101 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="068a6ab1-1a57-489b-adf2-67cb4011216e" containerName="keystone-bootstrap" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.399409 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="e786364e-fd43-43ed-a371-b87cc5a1e59a" containerName="dnsmasq-dns" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.399440 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="068a6ab1-1a57-489b-adf2-67cb4011216e" containerName="keystone-bootstrap" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.400564 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.403673 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.404377 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.404745 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.404887 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8kg7s" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.406788 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.407042 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.418164 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-758b4b5558-442kz"] Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.471598 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-combined-ca-bundle\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.471655 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6pxz\" (UniqueName: \"kubernetes.io/projected/883df2ca-c841-4882-b375-c74f7bf108f1-kube-api-access-k6pxz\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.471759 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-config-data\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.471838 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-public-tls-certs\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.471860 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-credential-keys\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.471886 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-fernet-keys\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.471928 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-internal-tls-certs\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.471956 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-scripts\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.574337 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6pxz\" (UniqueName: \"kubernetes.io/projected/883df2ca-c841-4882-b375-c74f7bf108f1-kube-api-access-k6pxz\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.574505 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-config-data\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.574649 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-credential-keys\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.574677 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-public-tls-certs\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.574707 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-fernet-keys\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.575657 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-internal-tls-certs\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.575687 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-scripts\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.575797 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-combined-ca-bundle\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.578887 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-credential-keys\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.579027 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-public-tls-certs\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.579500 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-config-data\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.579889 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-fernet-keys\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.580569 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-combined-ca-bundle\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.580584 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-internal-tls-certs\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.592583 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/883df2ca-c841-4882-b375-c74f7bf108f1-scripts\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.593810 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6pxz\" (UniqueName: \"kubernetes.io/projected/883df2ca-c841-4882-b375-c74f7bf108f1-kube-api-access-k6pxz\") pod \"keystone-758b4b5558-442kz\" (UID: \"883df2ca-c841-4882-b375-c74f7bf108f1\") " pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:57 crc kubenswrapper[4771]: I1002 10:00:57.738666 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:58 crc kubenswrapper[4771]: I1002 10:00:58.337191 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-758b4b5558-442kz"] Oct 02 10:00:59 crc kubenswrapper[4771]: I1002 10:00:59.289908 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-758b4b5558-442kz" event={"ID":"883df2ca-c841-4882-b375-c74f7bf108f1","Type":"ContainerStarted","Data":"fb9a15afa7c46e4e6feb018d5595f27918b322e86bfd4ae113ffe4667672de58"} Oct 02 10:00:59 crc kubenswrapper[4771]: I1002 10:00:59.290451 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-758b4b5558-442kz" event={"ID":"883df2ca-c841-4882-b375-c74f7bf108f1","Type":"ContainerStarted","Data":"5b12e24cfca7c6655054fa2126d92aa9737e945f9ae7e2a41e497ff743189fc5"} Oct 02 10:00:59 crc kubenswrapper[4771]: I1002 10:00:59.290469 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:00:59 crc kubenswrapper[4771]: I1002 10:00:59.320609 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-758b4b5558-442kz" podStartSLOduration=2.32058946 podStartE2EDuration="2.32058946s" podCreationTimestamp="2025-10-02 10:00:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:59.319007927 +0000 UTC m=+1446.966692994" watchObservedRunningTime="2025-10-02 10:00:59.32058946 +0000 UTC m=+1446.968274527" Oct 02 10:00:59 crc kubenswrapper[4771]: I1002 10:00:59.424740 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 10:00:59 crc kubenswrapper[4771]: I1002 10:00:59.424859 4771 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 10:00:59 crc kubenswrapper[4771]: I1002 10:00:59.428656 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 10:00:59 crc kubenswrapper[4771]: I1002 10:00:59.466343 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 10:00:59 crc kubenswrapper[4771]: I1002 10:00:59.466524 4771 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 10:00:59 crc kubenswrapper[4771]: I1002 10:00:59.470229 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.222270 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29323321-8rm5d"] Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.225604 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.245249 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323321-8rm5d"] Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.302999 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq5v9\" (UniqueName: \"kubernetes.io/projected/c8986807-60cd-43ec-83f3-f9339b28c16f-kube-api-access-vq5v9\") pod \"keystone-cron-29323321-8rm5d\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.303287 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-fernet-keys\") pod \"keystone-cron-29323321-8rm5d\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.303343 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-config-data\") pod \"keystone-cron-29323321-8rm5d\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.303363 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-combined-ca-bundle\") pod \"keystone-cron-29323321-8rm5d\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.314587 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5dxhb" event={"ID":"3bc64084-015e-47d2-b20c-f241e367fe27","Type":"ContainerStarted","Data":"0acd8c30b93bd4b48f322f37ba31cfee0224365b213479f44020970df277122b"} Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.338082 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-5dxhb" podStartSLOduration=5.322397931 podStartE2EDuration="48.338059444s" podCreationTimestamp="2025-10-02 10:00:12 +0000 UTC" firstStartedPulling="2025-10-02 10:00:15.769244688 +0000 UTC m=+1403.416929755" lastFinishedPulling="2025-10-02 10:00:58.784906201 +0000 UTC m=+1446.432591268" observedRunningTime="2025-10-02 10:01:00.33346458 +0000 UTC m=+1447.981149647" watchObservedRunningTime="2025-10-02 10:01:00.338059444 +0000 UTC m=+1447.985744511" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.405516 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-config-data\") pod \"keystone-cron-29323321-8rm5d\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.405567 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-combined-ca-bundle\") pod \"keystone-cron-29323321-8rm5d\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.405767 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq5v9\" (UniqueName: \"kubernetes.io/projected/c8986807-60cd-43ec-83f3-f9339b28c16f-kube-api-access-vq5v9\") pod \"keystone-cron-29323321-8rm5d\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.405977 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-fernet-keys\") pod \"keystone-cron-29323321-8rm5d\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.411776 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-combined-ca-bundle\") pod \"keystone-cron-29323321-8rm5d\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.416826 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-config-data\") pod \"keystone-cron-29323321-8rm5d\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.416852 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-fernet-keys\") pod \"keystone-cron-29323321-8rm5d\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.424310 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq5v9\" (UniqueName: \"kubernetes.io/projected/c8986807-60cd-43ec-83f3-f9339b28c16f-kube-api-access-vq5v9\") pod \"keystone-cron-29323321-8rm5d\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:00 crc kubenswrapper[4771]: I1002 10:01:00.565774 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:01 crc kubenswrapper[4771]: I1002 10:01:01.092538 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j4c5j" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="registry-server" probeResult="failure" output=< Oct 02 10:01:01 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:01:01 crc kubenswrapper[4771]: > Oct 02 10:01:01 crc kubenswrapper[4771]: I1002 10:01:01.203091 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323321-8rm5d"] Oct 02 10:01:01 crc kubenswrapper[4771]: I1002 10:01:01.377111 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wg2x2" event={"ID":"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5","Type":"ContainerStarted","Data":"8badc56bf937b29d49b7199b137adf1a05ea52645f9884aff25076cc4a12b257"} Oct 02 10:01:01 crc kubenswrapper[4771]: I1002 10:01:01.385385 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pkkwb" event={"ID":"2a276f13-9f48-40ae-8445-ce376fce62a1","Type":"ContainerStarted","Data":"f001212e1a5ae1a73fad0563d8f9c34c580f6d46c6321e4e686f1062692a8922"} Oct 02 10:01:01 crc kubenswrapper[4771]: I1002 10:01:01.405760 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323321-8rm5d" event={"ID":"c8986807-60cd-43ec-83f3-f9339b28c16f","Type":"ContainerStarted","Data":"f030ed13be5f39f4dc6f169c189f8d2f3669c01ab6e6703d945e8a4cdecf8b2d"} Oct 02 10:01:01 crc kubenswrapper[4771]: I1002 10:01:01.409708 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-wg2x2" podStartSLOduration=4.402384137 podStartE2EDuration="49.409678798s" podCreationTimestamp="2025-10-02 10:00:12 +0000 UTC" firstStartedPulling="2025-10-02 10:00:15.809780514 +0000 UTC m=+1403.457465581" lastFinishedPulling="2025-10-02 10:01:00.817075155 +0000 UTC m=+1448.464760242" observedRunningTime="2025-10-02 10:01:01.396243636 +0000 UTC m=+1449.043928723" watchObservedRunningTime="2025-10-02 10:01:01.409678798 +0000 UTC m=+1449.057363865" Oct 02 10:01:01 crc kubenswrapper[4771]: I1002 10:01:01.428951 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-pkkwb" podStartSLOduration=4.784749553 podStartE2EDuration="49.428906457s" podCreationTimestamp="2025-10-02 10:00:12 +0000 UTC" firstStartedPulling="2025-10-02 10:00:15.232080992 +0000 UTC m=+1402.879766059" lastFinishedPulling="2025-10-02 10:00:59.876237896 +0000 UTC m=+1447.523922963" observedRunningTime="2025-10-02 10:01:01.418365252 +0000 UTC m=+1449.066050319" watchObservedRunningTime="2025-10-02 10:01:01.428906457 +0000 UTC m=+1449.076591534" Oct 02 10:01:02 crc kubenswrapper[4771]: I1002 10:01:02.442000 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323321-8rm5d" event={"ID":"c8986807-60cd-43ec-83f3-f9339b28c16f","Type":"ContainerStarted","Data":"ba568df123ad468283e873b0452f5463ea7dd6aa4a576c7692d45eed93ca6cfd"} Oct 02 10:01:02 crc kubenswrapper[4771]: I1002 10:01:02.469000 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29323321-8rm5d" podStartSLOduration=2.46896033 podStartE2EDuration="2.46896033s" podCreationTimestamp="2025-10-02 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:02.468433966 +0000 UTC m=+1450.116119043" watchObservedRunningTime="2025-10-02 10:01:02.46896033 +0000 UTC m=+1450.116645397" Oct 02 10:01:03 crc kubenswrapper[4771]: I1002 10:01:03.799753 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jmbf9"] Oct 02 10:01:03 crc kubenswrapper[4771]: I1002 10:01:03.804955 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:03 crc kubenswrapper[4771]: I1002 10:01:03.817054 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jmbf9"] Oct 02 10:01:03 crc kubenswrapper[4771]: I1002 10:01:03.907269 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47fce250-07d5-40e4-9a1d-009285bc460f-utilities\") pod \"certified-operators-jmbf9\" (UID: \"47fce250-07d5-40e4-9a1d-009285bc460f\") " pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:03 crc kubenswrapper[4771]: I1002 10:01:03.907387 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47fce250-07d5-40e4-9a1d-009285bc460f-catalog-content\") pod \"certified-operators-jmbf9\" (UID: \"47fce250-07d5-40e4-9a1d-009285bc460f\") " pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:03 crc kubenswrapper[4771]: I1002 10:01:03.907528 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjf48\" (UniqueName: \"kubernetes.io/projected/47fce250-07d5-40e4-9a1d-009285bc460f-kube-api-access-hjf48\") pod \"certified-operators-jmbf9\" (UID: \"47fce250-07d5-40e4-9a1d-009285bc460f\") " pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:04 crc kubenswrapper[4771]: I1002 10:01:04.009777 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjf48\" (UniqueName: \"kubernetes.io/projected/47fce250-07d5-40e4-9a1d-009285bc460f-kube-api-access-hjf48\") pod \"certified-operators-jmbf9\" (UID: \"47fce250-07d5-40e4-9a1d-009285bc460f\") " pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:04 crc kubenswrapper[4771]: I1002 10:01:04.009880 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47fce250-07d5-40e4-9a1d-009285bc460f-utilities\") pod \"certified-operators-jmbf9\" (UID: \"47fce250-07d5-40e4-9a1d-009285bc460f\") " pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:04 crc kubenswrapper[4771]: I1002 10:01:04.009933 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47fce250-07d5-40e4-9a1d-009285bc460f-catalog-content\") pod \"certified-operators-jmbf9\" (UID: \"47fce250-07d5-40e4-9a1d-009285bc460f\") " pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:04 crc kubenswrapper[4771]: I1002 10:01:04.010550 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47fce250-07d5-40e4-9a1d-009285bc460f-catalog-content\") pod \"certified-operators-jmbf9\" (UID: \"47fce250-07d5-40e4-9a1d-009285bc460f\") " pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:04 crc kubenswrapper[4771]: I1002 10:01:04.011929 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47fce250-07d5-40e4-9a1d-009285bc460f-utilities\") pod \"certified-operators-jmbf9\" (UID: \"47fce250-07d5-40e4-9a1d-009285bc460f\") " pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:04 crc kubenswrapper[4771]: I1002 10:01:04.044632 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjf48\" (UniqueName: \"kubernetes.io/projected/47fce250-07d5-40e4-9a1d-009285bc460f-kube-api-access-hjf48\") pod \"certified-operators-jmbf9\" (UID: \"47fce250-07d5-40e4-9a1d-009285bc460f\") " pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:04 crc kubenswrapper[4771]: I1002 10:01:04.188267 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:05 crc kubenswrapper[4771]: I1002 10:01:05.481528 4771 generic.go:334] "Generic (PLEG): container finished" podID="c8986807-60cd-43ec-83f3-f9339b28c16f" containerID="ba568df123ad468283e873b0452f5463ea7dd6aa4a576c7692d45eed93ca6cfd" exitCode=0 Oct 02 10:01:05 crc kubenswrapper[4771]: I1002 10:01:05.481577 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323321-8rm5d" event={"ID":"c8986807-60cd-43ec-83f3-f9339b28c16f","Type":"ContainerDied","Data":"ba568df123ad468283e873b0452f5463ea7dd6aa4a576c7692d45eed93ca6cfd"} Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.404268 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.504795 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-config-data\") pod \"c8986807-60cd-43ec-83f3-f9339b28c16f\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.504865 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-combined-ca-bundle\") pod \"c8986807-60cd-43ec-83f3-f9339b28c16f\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.504990 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-fernet-keys\") pod \"c8986807-60cd-43ec-83f3-f9339b28c16f\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.505304 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq5v9\" (UniqueName: \"kubernetes.io/projected/c8986807-60cd-43ec-83f3-f9339b28c16f-kube-api-access-vq5v9\") pod \"c8986807-60cd-43ec-83f3-f9339b28c16f\" (UID: \"c8986807-60cd-43ec-83f3-f9339b28c16f\") " Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.520289 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323321-8rm5d" event={"ID":"c8986807-60cd-43ec-83f3-f9339b28c16f","Type":"ContainerDied","Data":"f030ed13be5f39f4dc6f169c189f8d2f3669c01ab6e6703d945e8a4cdecf8b2d"} Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.520347 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f030ed13be5f39f4dc6f169c189f8d2f3669c01ab6e6703d945e8a4cdecf8b2d" Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.520353 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323321-8rm5d" Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.528777 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8986807-60cd-43ec-83f3-f9339b28c16f-kube-api-access-vq5v9" (OuterVolumeSpecName: "kube-api-access-vq5v9") pod "c8986807-60cd-43ec-83f3-f9339b28c16f" (UID: "c8986807-60cd-43ec-83f3-f9339b28c16f"). InnerVolumeSpecName "kube-api-access-vq5v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.530412 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c8986807-60cd-43ec-83f3-f9339b28c16f" (UID: "c8986807-60cd-43ec-83f3-f9339b28c16f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.553599 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8986807-60cd-43ec-83f3-f9339b28c16f" (UID: "c8986807-60cd-43ec-83f3-f9339b28c16f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.600834 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-config-data" (OuterVolumeSpecName: "config-data") pod "c8986807-60cd-43ec-83f3-f9339b28c16f" (UID: "c8986807-60cd-43ec-83f3-f9339b28c16f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.608808 4771 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.608857 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq5v9\" (UniqueName: \"kubernetes.io/projected/c8986807-60cd-43ec-83f3-f9339b28c16f-kube-api-access-vq5v9\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.608872 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:07 crc kubenswrapper[4771]: I1002 10:01:07.608883 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8986807-60cd-43ec-83f3-f9339b28c16f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:08 crc kubenswrapper[4771]: I1002 10:01:08.561880 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jmbf9"] Oct 02 10:01:08 crc kubenswrapper[4771]: W1002 10:01:08.570669 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47fce250_07d5_40e4_9a1d_009285bc460f.slice/crio-71eafb015a9e4a384843e8ffca77e869b4f0825753ca5c3deaf575bcace514b4 WatchSource:0}: Error finding container 71eafb015a9e4a384843e8ffca77e869b4f0825753ca5c3deaf575bcace514b4: Status 404 returned error can't find the container with id 71eafb015a9e4a384843e8ffca77e869b4f0825753ca5c3deaf575bcace514b4 Oct 02 10:01:09 crc kubenswrapper[4771]: I1002 10:01:09.542862 4771 generic.go:334] "Generic (PLEG): container finished" podID="47fce250-07d5-40e4-9a1d-009285bc460f" containerID="193d5055706cf3d956a16df74de9a60d2e1547f11d36babf2d507b3cde15055e" exitCode=0 Oct 02 10:01:09 crc kubenswrapper[4771]: I1002 10:01:09.542963 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jmbf9" event={"ID":"47fce250-07d5-40e4-9a1d-009285bc460f","Type":"ContainerDied","Data":"193d5055706cf3d956a16df74de9a60d2e1547f11d36babf2d507b3cde15055e"} Oct 02 10:01:09 crc kubenswrapper[4771]: I1002 10:01:09.543281 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jmbf9" event={"ID":"47fce250-07d5-40e4-9a1d-009285bc460f","Type":"ContainerStarted","Data":"71eafb015a9e4a384843e8ffca77e869b4f0825753ca5c3deaf575bcace514b4"} Oct 02 10:01:11 crc kubenswrapper[4771]: I1002 10:01:11.111398 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j4c5j" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="registry-server" probeResult="failure" output=< Oct 02 10:01:11 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:01:11 crc kubenswrapper[4771]: > Oct 02 10:01:11 crc kubenswrapper[4771]: I1002 10:01:11.571716 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"48fe997b-eed7-4191-9f63-ecb40112cf50","Type":"ContainerStarted","Data":"1981281e06d0afbaf2803e5d8e62a4b940d42eda3e53c16ba6319c58e2c89cfd"} Oct 02 10:01:11 crc kubenswrapper[4771]: I1002 10:01:11.571999 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="ceilometer-notification-agent" containerID="cri-o://32c91bd9bf80e1b4b42d19b0a1d5a478d0fe72de8bdfc148be56fb134cfdcf42" gracePeriod=30 Oct 02 10:01:11 crc kubenswrapper[4771]: I1002 10:01:11.571952 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="sg-core" containerID="cri-o://e9816a93a5fdd5b53097989567d5375baef8e01f9e272c6a9c166cb806660219" gracePeriod=30 Oct 02 10:01:11 crc kubenswrapper[4771]: I1002 10:01:11.572225 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:01:11 crc kubenswrapper[4771]: I1002 10:01:11.571928 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="ceilometer-central-agent" containerID="cri-o://db909e1281bba61267bf0467d67eb148eb35688311d23f6a4c31582d35f73211" gracePeriod=30 Oct 02 10:01:11 crc kubenswrapper[4771]: I1002 10:01:11.572002 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="proxy-httpd" containerID="cri-o://1981281e06d0afbaf2803e5d8e62a4b940d42eda3e53c16ba6319c58e2c89cfd" gracePeriod=30 Oct 02 10:01:11 crc kubenswrapper[4771]: I1002 10:01:11.606348 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.181276192 podStartE2EDuration="58.606320667s" podCreationTimestamp="2025-10-02 10:00:13 +0000 UTC" firstStartedPulling="2025-10-02 10:00:16.539117004 +0000 UTC m=+1404.186802071" lastFinishedPulling="2025-10-02 10:01:10.964161479 +0000 UTC m=+1458.611846546" observedRunningTime="2025-10-02 10:01:11.598101606 +0000 UTC m=+1459.245786673" watchObservedRunningTime="2025-10-02 10:01:11.606320667 +0000 UTC m=+1459.254005734" Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.587947 4771 generic.go:334] "Generic (PLEG): container finished" podID="47fce250-07d5-40e4-9a1d-009285bc460f" containerID="07ad202028d929b0d5c7ae153b631470ea628deb7298ef99dbb197215eb227b6" exitCode=0 Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.588063 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jmbf9" event={"ID":"47fce250-07d5-40e4-9a1d-009285bc460f","Type":"ContainerDied","Data":"07ad202028d929b0d5c7ae153b631470ea628deb7298ef99dbb197215eb227b6"} Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.604232 4771 generic.go:334] "Generic (PLEG): container finished" podID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerID="1981281e06d0afbaf2803e5d8e62a4b940d42eda3e53c16ba6319c58e2c89cfd" exitCode=0 Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.604278 4771 generic.go:334] "Generic (PLEG): container finished" podID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerID="e9816a93a5fdd5b53097989567d5375baef8e01f9e272c6a9c166cb806660219" exitCode=2 Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.604288 4771 generic.go:334] "Generic (PLEG): container finished" podID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerID="32c91bd9bf80e1b4b42d19b0a1d5a478d0fe72de8bdfc148be56fb134cfdcf42" exitCode=0 Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.604297 4771 generic.go:334] "Generic (PLEG): container finished" podID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerID="db909e1281bba61267bf0467d67eb148eb35688311d23f6a4c31582d35f73211" exitCode=0 Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.604337 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"48fe997b-eed7-4191-9f63-ecb40112cf50","Type":"ContainerDied","Data":"1981281e06d0afbaf2803e5d8e62a4b940d42eda3e53c16ba6319c58e2c89cfd"} Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.604414 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"48fe997b-eed7-4191-9f63-ecb40112cf50","Type":"ContainerDied","Data":"e9816a93a5fdd5b53097989567d5375baef8e01f9e272c6a9c166cb806660219"} Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.604428 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"48fe997b-eed7-4191-9f63-ecb40112cf50","Type":"ContainerDied","Data":"32c91bd9bf80e1b4b42d19b0a1d5a478d0fe72de8bdfc148be56fb134cfdcf42"} Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.604439 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"48fe997b-eed7-4191-9f63-ecb40112cf50","Type":"ContainerDied","Data":"db909e1281bba61267bf0467d67eb148eb35688311d23f6a4c31582d35f73211"} Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.925392 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.944049 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-scripts\") pod \"48fe997b-eed7-4191-9f63-ecb40112cf50\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.944105 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-combined-ca-bundle\") pod \"48fe997b-eed7-4191-9f63-ecb40112cf50\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.944157 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-sg-core-conf-yaml\") pod \"48fe997b-eed7-4191-9f63-ecb40112cf50\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.944176 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48fe997b-eed7-4191-9f63-ecb40112cf50-run-httpd\") pod \"48fe997b-eed7-4191-9f63-ecb40112cf50\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.944216 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-config-data\") pod \"48fe997b-eed7-4191-9f63-ecb40112cf50\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.944327 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48fe997b-eed7-4191-9f63-ecb40112cf50-log-httpd\") pod \"48fe997b-eed7-4191-9f63-ecb40112cf50\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.944382 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jzxj\" (UniqueName: \"kubernetes.io/projected/48fe997b-eed7-4191-9f63-ecb40112cf50-kube-api-access-9jzxj\") pod \"48fe997b-eed7-4191-9f63-ecb40112cf50\" (UID: \"48fe997b-eed7-4191-9f63-ecb40112cf50\") " Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.945634 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48fe997b-eed7-4191-9f63-ecb40112cf50-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "48fe997b-eed7-4191-9f63-ecb40112cf50" (UID: "48fe997b-eed7-4191-9f63-ecb40112cf50"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.945735 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48fe997b-eed7-4191-9f63-ecb40112cf50-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "48fe997b-eed7-4191-9f63-ecb40112cf50" (UID: "48fe997b-eed7-4191-9f63-ecb40112cf50"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.955463 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-scripts" (OuterVolumeSpecName: "scripts") pod "48fe997b-eed7-4191-9f63-ecb40112cf50" (UID: "48fe997b-eed7-4191-9f63-ecb40112cf50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:12 crc kubenswrapper[4771]: I1002 10:01:12.965160 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48fe997b-eed7-4191-9f63-ecb40112cf50-kube-api-access-9jzxj" (OuterVolumeSpecName: "kube-api-access-9jzxj") pod "48fe997b-eed7-4191-9f63-ecb40112cf50" (UID: "48fe997b-eed7-4191-9f63-ecb40112cf50"). InnerVolumeSpecName "kube-api-access-9jzxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.011972 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "48fe997b-eed7-4191-9f63-ecb40112cf50" (UID: "48fe997b-eed7-4191-9f63-ecb40112cf50"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.047146 4771 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48fe997b-eed7-4191-9f63-ecb40112cf50-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.047256 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jzxj\" (UniqueName: \"kubernetes.io/projected/48fe997b-eed7-4191-9f63-ecb40112cf50-kube-api-access-9jzxj\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.047270 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.047278 4771 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48fe997b-eed7-4191-9f63-ecb40112cf50-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.047290 4771 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.064044 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48fe997b-eed7-4191-9f63-ecb40112cf50" (UID: "48fe997b-eed7-4191-9f63-ecb40112cf50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.097292 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-config-data" (OuterVolumeSpecName: "config-data") pod "48fe997b-eed7-4191-9f63-ecb40112cf50" (UID: "48fe997b-eed7-4191-9f63-ecb40112cf50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.149587 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.149882 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48fe997b-eed7-4191-9f63-ecb40112cf50-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.618313 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jmbf9" event={"ID":"47fce250-07d5-40e4-9a1d-009285bc460f","Type":"ContainerStarted","Data":"0c08b7daaa410e763a98ebf7c87dc9ea59f9163c7d40425a16028cecd7721502"} Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.621406 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"48fe997b-eed7-4191-9f63-ecb40112cf50","Type":"ContainerDied","Data":"a8b6e5a6541427a6b5651096dd99c42304ee54dd6f7da92fa12b9589e3b5e705"} Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.621443 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.621465 4771 scope.go:117] "RemoveContainer" containerID="1981281e06d0afbaf2803e5d8e62a4b940d42eda3e53c16ba6319c58e2c89cfd" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.624686 4771 generic.go:334] "Generic (PLEG): container finished" podID="a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5" containerID="8badc56bf937b29d49b7199b137adf1a05ea52645f9884aff25076cc4a12b257" exitCode=0 Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.624741 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wg2x2" event={"ID":"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5","Type":"ContainerDied","Data":"8badc56bf937b29d49b7199b137adf1a05ea52645f9884aff25076cc4a12b257"} Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.650545 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jmbf9" podStartSLOduration=8.404047266 podStartE2EDuration="10.650523477s" podCreationTimestamp="2025-10-02 10:01:03 +0000 UTC" firstStartedPulling="2025-10-02 10:01:10.84027585 +0000 UTC m=+1458.487960917" lastFinishedPulling="2025-10-02 10:01:13.086752061 +0000 UTC m=+1460.734437128" observedRunningTime="2025-10-02 10:01:13.640811785 +0000 UTC m=+1461.288496852" watchObservedRunningTime="2025-10-02 10:01:13.650523477 +0000 UTC m=+1461.298208534" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.655296 4771 scope.go:117] "RemoveContainer" containerID="e9816a93a5fdd5b53097989567d5375baef8e01f9e272c6a9c166cb806660219" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.722611 4771 scope.go:117] "RemoveContainer" containerID="32c91bd9bf80e1b4b42d19b0a1d5a478d0fe72de8bdfc148be56fb134cfdcf42" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.797103 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.809662 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.817264 4771 scope.go:117] "RemoveContainer" containerID="db909e1281bba61267bf0467d67eb148eb35688311d23f6a4c31582d35f73211" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.831176 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:01:13 crc kubenswrapper[4771]: E1002 10:01:13.832302 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="ceilometer-central-agent" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.832409 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="ceilometer-central-agent" Oct 02 10:01:13 crc kubenswrapper[4771]: E1002 10:01:13.832540 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="ceilometer-notification-agent" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.832615 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="ceilometer-notification-agent" Oct 02 10:01:13 crc kubenswrapper[4771]: E1002 10:01:13.832763 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8986807-60cd-43ec-83f3-f9339b28c16f" containerName="keystone-cron" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.832841 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8986807-60cd-43ec-83f3-f9339b28c16f" containerName="keystone-cron" Oct 02 10:01:13 crc kubenswrapper[4771]: E1002 10:01:13.832947 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="sg-core" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.833028 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="sg-core" Oct 02 10:01:13 crc kubenswrapper[4771]: E1002 10:01:13.833108 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="proxy-httpd" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.833210 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="proxy-httpd" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.833582 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="sg-core" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.833685 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="proxy-httpd" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.833888 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="ceilometer-central-agent" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.833969 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8986807-60cd-43ec-83f3-f9339b28c16f" containerName="keystone-cron" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.834065 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" containerName="ceilometer-notification-agent" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.837389 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.845750 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.846882 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.868319 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.873334 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.874201 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-scripts\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.874287 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.874333 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-config-data\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.874961 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5ms5\" (UniqueName: \"kubernetes.io/projected/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-kube-api-access-q5ms5\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.875284 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-run-httpd\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.875339 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-log-httpd\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.977680 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5ms5\" (UniqueName: \"kubernetes.io/projected/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-kube-api-access-q5ms5\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.978145 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-run-httpd\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.978248 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-log-httpd\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.978406 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.978494 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-scripts\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.978593 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.978668 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-config-data\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.978591 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-run-httpd\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.979490 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-log-httpd\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.984604 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.985591 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-config-data\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.986085 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:13 crc kubenswrapper[4771]: I1002 10:01:13.988184 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-scripts\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:14 crc kubenswrapper[4771]: I1002 10:01:14.005845 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5ms5\" (UniqueName: \"kubernetes.io/projected/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-kube-api-access-q5ms5\") pod \"ceilometer-0\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " pod="openstack/ceilometer-0" Oct 02 10:01:14 crc kubenswrapper[4771]: I1002 10:01:14.189331 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:14 crc kubenswrapper[4771]: I1002 10:01:14.189407 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:14 crc kubenswrapper[4771]: I1002 10:01:14.197196 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:01:14 crc kubenswrapper[4771]: I1002 10:01:14.867308 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.214643 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wg2x2" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.250154 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jmbf9" podUID="47fce250-07d5-40e4-9a1d-009285bc460f" containerName="registry-server" probeResult="failure" output=< Oct 02 10:01:15 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:01:15 crc kubenswrapper[4771]: > Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.331911 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-scripts\") pod \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.332021 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khdp2\" (UniqueName: \"kubernetes.io/projected/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-kube-api-access-khdp2\") pod \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.332091 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-logs\") pod \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.332280 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-combined-ca-bundle\") pod \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.332379 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-config-data\") pod \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\" (UID: \"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5\") " Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.332804 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-logs" (OuterVolumeSpecName: "logs") pod "a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5" (UID: "a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.333444 4771 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.339046 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-kube-api-access-khdp2" (OuterVolumeSpecName: "kube-api-access-khdp2") pod "a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5" (UID: "a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5"). InnerVolumeSpecName "kube-api-access-khdp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.339375 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-scripts" (OuterVolumeSpecName: "scripts") pod "a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5" (UID: "a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.377773 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-config-data" (OuterVolumeSpecName: "config-data") pod "a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5" (UID: "a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.380000 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5" (UID: "a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.436159 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.436197 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.436210 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.436220 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khdp2\" (UniqueName: \"kubernetes.io/projected/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5-kube-api-access-khdp2\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.662275 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f","Type":"ContainerStarted","Data":"86964390f1dd2ff080457f91a091dc8c92b2f3352e7a275c41b0c493c9f150ca"} Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.664904 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wg2x2" event={"ID":"a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5","Type":"ContainerDied","Data":"ab59f9c82e7bce4474b8df1019c778ed326d14033177dc57bc9d0da16830f8f0"} Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.664961 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab59f9c82e7bce4474b8df1019c778ed326d14033177dc57bc9d0da16830f8f0" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.664981 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wg2x2" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.720191 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48fe997b-eed7-4191-9f63-ecb40112cf50" path="/var/lib/kubelet/pods/48fe997b-eed7-4191-9f63-ecb40112cf50/volumes" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.864429 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-58587f9544-7mwm6"] Oct 02 10:01:15 crc kubenswrapper[4771]: E1002 10:01:15.865988 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5" containerName="placement-db-sync" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.866088 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5" containerName="placement-db-sync" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.866889 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5" containerName="placement-db-sync" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.872200 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.878546 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.879379 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.879477 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.879555 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.882799 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-fmxrj" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.932946 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-58587f9544-7mwm6"] Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.962999 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd5227df-e476-4c2d-9ec2-694ab19dcb78-logs\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.963090 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-scripts\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.963175 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-public-tls-certs\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.963575 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6bgf\" (UniqueName: \"kubernetes.io/projected/bd5227df-e476-4c2d-9ec2-694ab19dcb78-kube-api-access-v6bgf\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.963644 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-combined-ca-bundle\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.963675 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-config-data\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:15 crc kubenswrapper[4771]: I1002 10:01:15.963925 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-internal-tls-certs\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.066217 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6bgf\" (UniqueName: \"kubernetes.io/projected/bd5227df-e476-4c2d-9ec2-694ab19dcb78-kube-api-access-v6bgf\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.066634 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-combined-ca-bundle\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.066676 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-config-data\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.066755 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-internal-tls-certs\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.066856 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd5227df-e476-4c2d-9ec2-694ab19dcb78-logs\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.066929 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-scripts\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.067026 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-public-tls-certs\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.067431 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd5227df-e476-4c2d-9ec2-694ab19dcb78-logs\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.071657 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-config-data\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.071760 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-public-tls-certs\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.072094 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-internal-tls-certs\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.072754 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-combined-ca-bundle\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.072995 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd5227df-e476-4c2d-9ec2-694ab19dcb78-scripts\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.088849 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6bgf\" (UniqueName: \"kubernetes.io/projected/bd5227df-e476-4c2d-9ec2-694ab19dcb78-kube-api-access-v6bgf\") pod \"placement-58587f9544-7mwm6\" (UID: \"bd5227df-e476-4c2d-9ec2-694ab19dcb78\") " pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.208138 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.680614 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f","Type":"ContainerStarted","Data":"74bf1bb5db2b07de6f282f2c7a2c8ac004d12d5baae3d4580a58192f6649d863"} Oct 02 10:01:16 crc kubenswrapper[4771]: I1002 10:01:16.715168 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-58587f9544-7mwm6"] Oct 02 10:01:16 crc kubenswrapper[4771]: W1002 10:01:16.719332 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd5227df_e476_4c2d_9ec2_694ab19dcb78.slice/crio-1abe50e24d6abae12b8be926b4dc96b42fc7f9e2f0d7b19c6736e27a92eafe03 WatchSource:0}: Error finding container 1abe50e24d6abae12b8be926b4dc96b42fc7f9e2f0d7b19c6736e27a92eafe03: Status 404 returned error can't find the container with id 1abe50e24d6abae12b8be926b4dc96b42fc7f9e2f0d7b19c6736e27a92eafe03 Oct 02 10:01:17 crc kubenswrapper[4771]: I1002 10:01:17.713946 4771 generic.go:334] "Generic (PLEG): container finished" podID="3bc64084-015e-47d2-b20c-f241e367fe27" containerID="0acd8c30b93bd4b48f322f37ba31cfee0224365b213479f44020970df277122b" exitCode=0 Oct 02 10:01:17 crc kubenswrapper[4771]: I1002 10:01:17.714052 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5dxhb" event={"ID":"3bc64084-015e-47d2-b20c-f241e367fe27","Type":"ContainerDied","Data":"0acd8c30b93bd4b48f322f37ba31cfee0224365b213479f44020970df277122b"} Oct 02 10:01:17 crc kubenswrapper[4771]: I1002 10:01:17.738842 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58587f9544-7mwm6" event={"ID":"bd5227df-e476-4c2d-9ec2-694ab19dcb78","Type":"ContainerStarted","Data":"68475d3ed116f5bbda0eda43e2c5f48371f4062db7da6db41d849b63d2925317"} Oct 02 10:01:17 crc kubenswrapper[4771]: I1002 10:01:17.738952 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58587f9544-7mwm6" event={"ID":"bd5227df-e476-4c2d-9ec2-694ab19dcb78","Type":"ContainerStarted","Data":"cc2f98869f2c58346a0cb6270dd92121715566468a6abe3a476d2b4fe5e03080"} Oct 02 10:01:17 crc kubenswrapper[4771]: I1002 10:01:17.738968 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58587f9544-7mwm6" event={"ID":"bd5227df-e476-4c2d-9ec2-694ab19dcb78","Type":"ContainerStarted","Data":"1abe50e24d6abae12b8be926b4dc96b42fc7f9e2f0d7b19c6736e27a92eafe03"} Oct 02 10:01:17 crc kubenswrapper[4771]: I1002 10:01:17.739010 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:17 crc kubenswrapper[4771]: I1002 10:01:17.739074 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:17 crc kubenswrapper[4771]: I1002 10:01:17.744470 4771 generic.go:334] "Generic (PLEG): container finished" podID="042a2fb4-504c-4448-922b-48db5507e964" containerID="e0a3421d25863823e2518b80b0336067bf482596d694069ade7893462313fb3a" exitCode=0 Oct 02 10:01:17 crc kubenswrapper[4771]: I1002 10:01:17.744522 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-dbght" event={"ID":"042a2fb4-504c-4448-922b-48db5507e964","Type":"ContainerDied","Data":"e0a3421d25863823e2518b80b0336067bf482596d694069ade7893462313fb3a"} Oct 02 10:01:17 crc kubenswrapper[4771]: I1002 10:01:17.760435 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f","Type":"ContainerStarted","Data":"00e4574206a45d648bcda5196849131bf2c3fb95fbf0397513229cc9a789de07"} Oct 02 10:01:17 crc kubenswrapper[4771]: I1002 10:01:17.789905 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-58587f9544-7mwm6" podStartSLOduration=2.789879909 podStartE2EDuration="2.789879909s" podCreationTimestamp="2025-10-02 10:01:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:17.778761909 +0000 UTC m=+1465.426446996" watchObservedRunningTime="2025-10-02 10:01:17.789879909 +0000 UTC m=+1465.437564976" Oct 02 10:01:18 crc kubenswrapper[4771]: I1002 10:01:18.785757 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f","Type":"ContainerStarted","Data":"48ffa7722cc5a44891b25a0bf45bdb35e6a21d55b848def01947b2101bd08f22"} Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.469618 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-dbght" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.478704 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5dxhb" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.568773 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wv5l\" (UniqueName: \"kubernetes.io/projected/3bc64084-015e-47d2-b20c-f241e367fe27-kube-api-access-9wv5l\") pod \"3bc64084-015e-47d2-b20c-f241e367fe27\" (UID: \"3bc64084-015e-47d2-b20c-f241e367fe27\") " Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.568962 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d56n\" (UniqueName: \"kubernetes.io/projected/042a2fb4-504c-4448-922b-48db5507e964-kube-api-access-8d56n\") pod \"042a2fb4-504c-4448-922b-48db5507e964\" (UID: \"042a2fb4-504c-4448-922b-48db5507e964\") " Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.569144 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc64084-015e-47d2-b20c-f241e367fe27-combined-ca-bundle\") pod \"3bc64084-015e-47d2-b20c-f241e367fe27\" (UID: \"3bc64084-015e-47d2-b20c-f241e367fe27\") " Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.569189 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bc64084-015e-47d2-b20c-f241e367fe27-db-sync-config-data\") pod \"3bc64084-015e-47d2-b20c-f241e367fe27\" (UID: \"3bc64084-015e-47d2-b20c-f241e367fe27\") " Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.569260 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042a2fb4-504c-4448-922b-48db5507e964-config-data\") pod \"042a2fb4-504c-4448-922b-48db5507e964\" (UID: \"042a2fb4-504c-4448-922b-48db5507e964\") " Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.569301 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042a2fb4-504c-4448-922b-48db5507e964-combined-ca-bundle\") pod \"042a2fb4-504c-4448-922b-48db5507e964\" (UID: \"042a2fb4-504c-4448-922b-48db5507e964\") " Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.616121 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bc64084-015e-47d2-b20c-f241e367fe27-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3bc64084-015e-47d2-b20c-f241e367fe27" (UID: "3bc64084-015e-47d2-b20c-f241e367fe27"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.619091 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/042a2fb4-504c-4448-922b-48db5507e964-kube-api-access-8d56n" (OuterVolumeSpecName: "kube-api-access-8d56n") pod "042a2fb4-504c-4448-922b-48db5507e964" (UID: "042a2fb4-504c-4448-922b-48db5507e964"). InnerVolumeSpecName "kube-api-access-8d56n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.646599 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bc64084-015e-47d2-b20c-f241e367fe27-kube-api-access-9wv5l" (OuterVolumeSpecName: "kube-api-access-9wv5l") pod "3bc64084-015e-47d2-b20c-f241e367fe27" (UID: "3bc64084-015e-47d2-b20c-f241e367fe27"). InnerVolumeSpecName "kube-api-access-9wv5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.674524 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/042a2fb4-504c-4448-922b-48db5507e964-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "042a2fb4-504c-4448-922b-48db5507e964" (UID: "042a2fb4-504c-4448-922b-48db5507e964"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.686513 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d56n\" (UniqueName: \"kubernetes.io/projected/042a2fb4-504c-4448-922b-48db5507e964-kube-api-access-8d56n\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.686844 4771 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bc64084-015e-47d2-b20c-f241e367fe27-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.686931 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042a2fb4-504c-4448-922b-48db5507e964-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.687009 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wv5l\" (UniqueName: \"kubernetes.io/projected/3bc64084-015e-47d2-b20c-f241e367fe27-kube-api-access-9wv5l\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.759660 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/042a2fb4-504c-4448-922b-48db5507e964-config-data" (OuterVolumeSpecName: "config-data") pod "042a2fb4-504c-4448-922b-48db5507e964" (UID: "042a2fb4-504c-4448-922b-48db5507e964"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.793460 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bc64084-015e-47d2-b20c-f241e367fe27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3bc64084-015e-47d2-b20c-f241e367fe27" (UID: "3bc64084-015e-47d2-b20c-f241e367fe27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.794929 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc64084-015e-47d2-b20c-f241e367fe27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.794955 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042a2fb4-504c-4448-922b-48db5507e964-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.821949 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5dxhb" event={"ID":"3bc64084-015e-47d2-b20c-f241e367fe27","Type":"ContainerDied","Data":"927fb362f2abeb29710ba042c5433d3f117f9e39673362f0c8626d4ea12253e3"} Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.822403 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="927fb362f2abeb29710ba042c5433d3f117f9e39673362f0c8626d4ea12253e3" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.822446 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5dxhb" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.830541 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-dbght" event={"ID":"042a2fb4-504c-4448-922b-48db5507e964","Type":"ContainerDied","Data":"8981d7e4955cdf2abed1091372d0858fa931528cd3cd9192d1f96a9fa2282b82"} Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.830599 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8981d7e4955cdf2abed1091372d0858fa931528cd3cd9192d1f96a9fa2282b82" Oct 02 10:01:19 crc kubenswrapper[4771]: I1002 10:01:19.830690 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-dbght" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.097430 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-64df995667-zkf22"] Oct 02 10:01:20 crc kubenswrapper[4771]: E1002 10:01:20.098471 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bc64084-015e-47d2-b20c-f241e367fe27" containerName="barbican-db-sync" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.098499 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bc64084-015e-47d2-b20c-f241e367fe27" containerName="barbican-db-sync" Oct 02 10:01:20 crc kubenswrapper[4771]: E1002 10:01:20.098547 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="042a2fb4-504c-4448-922b-48db5507e964" containerName="heat-db-sync" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.098556 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="042a2fb4-504c-4448-922b-48db5507e964" containerName="heat-db-sync" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.098880 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bc64084-015e-47d2-b20c-f241e367fe27" containerName="barbican-db-sync" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.098911 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="042a2fb4-504c-4448-922b-48db5507e964" containerName="heat-db-sync" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.100581 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.107953 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-lqb76" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.108490 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.117711 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.128163 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-64df995667-zkf22"] Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.143003 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6998cd89db-l2xv7"] Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.146009 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.156241 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.181636 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6998cd89db-l2xv7"] Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.207418 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt5kc\" (UniqueName: \"kubernetes.io/projected/3439cba5-f86c-48f3-8d0b-587c02148975-kube-api-access-xt5kc\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.207485 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3439cba5-f86c-48f3-8d0b-587c02148975-logs\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.207518 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b69ba24b-3361-44d2-8774-64230ab7a376-config-data-custom\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.207643 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b69ba24b-3361-44d2-8774-64230ab7a376-logs\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.207680 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpw9m\" (UniqueName: \"kubernetes.io/projected/b69ba24b-3361-44d2-8774-64230ab7a376-kube-api-access-wpw9m\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.207708 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3439cba5-f86c-48f3-8d0b-587c02148975-config-data-custom\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.207746 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b69ba24b-3361-44d2-8774-64230ab7a376-combined-ca-bundle\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.207776 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b69ba24b-3361-44d2-8774-64230ab7a376-config-data\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.207808 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3439cba5-f86c-48f3-8d0b-587c02148975-config-data\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.207852 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3439cba5-f86c-48f3-8d0b-587c02148975-combined-ca-bundle\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.268089 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-hh2km"] Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.273271 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.295509 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-hh2km"] Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.311892 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-config\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312045 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b69ba24b-3361-44d2-8774-64230ab7a376-logs\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312091 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpw9m\" (UniqueName: \"kubernetes.io/projected/b69ba24b-3361-44d2-8774-64230ab7a376-kube-api-access-wpw9m\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312154 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3439cba5-f86c-48f3-8d0b-587c02148975-config-data-custom\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312201 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312250 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312288 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b69ba24b-3361-44d2-8774-64230ab7a376-combined-ca-bundle\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312340 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312376 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b69ba24b-3361-44d2-8774-64230ab7a376-config-data\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312396 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312453 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3439cba5-f86c-48f3-8d0b-587c02148975-config-data\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312534 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3439cba5-f86c-48f3-8d0b-587c02148975-combined-ca-bundle\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312605 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnztm\" (UniqueName: \"kubernetes.io/projected/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-kube-api-access-wnztm\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312636 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt5kc\" (UniqueName: \"kubernetes.io/projected/3439cba5-f86c-48f3-8d0b-587c02148975-kube-api-access-xt5kc\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312677 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3439cba5-f86c-48f3-8d0b-587c02148975-logs\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.312722 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b69ba24b-3361-44d2-8774-64230ab7a376-config-data-custom\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.313336 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b69ba24b-3361-44d2-8774-64230ab7a376-logs\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.319960 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3439cba5-f86c-48f3-8d0b-587c02148975-logs\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.333297 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b69ba24b-3361-44d2-8774-64230ab7a376-combined-ca-bundle\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.336510 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b69ba24b-3361-44d2-8774-64230ab7a376-config-data\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.336857 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b69ba24b-3361-44d2-8774-64230ab7a376-config-data-custom\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.337020 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3439cba5-f86c-48f3-8d0b-587c02148975-config-data-custom\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.337405 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3439cba5-f86c-48f3-8d0b-587c02148975-config-data\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.339151 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3439cba5-f86c-48f3-8d0b-587c02148975-combined-ca-bundle\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.352406 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpw9m\" (UniqueName: \"kubernetes.io/projected/b69ba24b-3361-44d2-8774-64230ab7a376-kube-api-access-wpw9m\") pod \"barbican-keystone-listener-6998cd89db-l2xv7\" (UID: \"b69ba24b-3361-44d2-8774-64230ab7a376\") " pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.354796 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt5kc\" (UniqueName: \"kubernetes.io/projected/3439cba5-f86c-48f3-8d0b-587c02148975-kube-api-access-xt5kc\") pod \"barbican-worker-64df995667-zkf22\" (UID: \"3439cba5-f86c-48f3-8d0b-587c02148975\") " pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.414855 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.414920 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.414953 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.415037 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnztm\" (UniqueName: \"kubernetes.io/projected/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-kube-api-access-wnztm\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.415111 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-config\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.415217 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.416318 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.416888 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.417412 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.417707 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-config\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.418308 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.438715 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-64df995667-zkf22" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.448984 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnztm\" (UniqueName: \"kubernetes.io/projected/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-kube-api-access-wnztm\") pod \"dnsmasq-dns-586bdc5f9-hh2km\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.487382 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5f7bdb69dd-8694c"] Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.489607 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.495200 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.513764 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.521075 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f7bdb69dd-8694c"] Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.597867 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.629340 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-config-data-custom\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.629524 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e17997e0-94d7-4a2f-83c1-dff9a23188cc-logs\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.629561 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-combined-ca-bundle\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.629608 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-config-data\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.629864 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htw9g\" (UniqueName: \"kubernetes.io/projected/e17997e0-94d7-4a2f-83c1-dff9a23188cc-kube-api-access-htw9g\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.733146 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e17997e0-94d7-4a2f-83c1-dff9a23188cc-logs\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.733189 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-combined-ca-bundle\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.733230 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-config-data\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.733459 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htw9g\" (UniqueName: \"kubernetes.io/projected/e17997e0-94d7-4a2f-83c1-dff9a23188cc-kube-api-access-htw9g\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.733537 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-config-data-custom\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.735909 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e17997e0-94d7-4a2f-83c1-dff9a23188cc-logs\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.745203 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-config-data-custom\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.745797 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-combined-ca-bundle\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.750441 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-config-data\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.774518 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htw9g\" (UniqueName: \"kubernetes.io/projected/e17997e0-94d7-4a2f-83c1-dff9a23188cc-kube-api-access-htw9g\") pod \"barbican-api-5f7bdb69dd-8694c\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.889887 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f","Type":"ContainerStarted","Data":"bb7ea1b59d47aa056d3f4df5e6ca203ac5091bfe041d713a62b00549177eb312"} Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.891813 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.924002 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.068276994 podStartE2EDuration="7.923983515s" podCreationTimestamp="2025-10-02 10:01:13 +0000 UTC" firstStartedPulling="2025-10-02 10:01:14.860644934 +0000 UTC m=+1462.508330001" lastFinishedPulling="2025-10-02 10:01:19.716351455 +0000 UTC m=+1467.364036522" observedRunningTime="2025-10-02 10:01:20.918779775 +0000 UTC m=+1468.566464842" watchObservedRunningTime="2025-10-02 10:01:20.923983515 +0000 UTC m=+1468.571668582" Oct 02 10:01:20 crc kubenswrapper[4771]: I1002 10:01:20.930945 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:21 crc kubenswrapper[4771]: I1002 10:01:21.141225 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j4c5j" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="registry-server" probeResult="failure" output=< Oct 02 10:01:21 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:01:21 crc kubenswrapper[4771]: > Oct 02 10:01:21 crc kubenswrapper[4771]: I1002 10:01:21.310301 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-64df995667-zkf22"] Oct 02 10:01:21 crc kubenswrapper[4771]: I1002 10:01:21.411769 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6998cd89db-l2xv7"] Oct 02 10:01:21 crc kubenswrapper[4771]: W1002 10:01:21.419483 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb69ba24b_3361_44d2_8774_64230ab7a376.slice/crio-bfe1fbf708356b7449f08faa913649f8a1790800bf293694ed55b2ce8e322c81 WatchSource:0}: Error finding container bfe1fbf708356b7449f08faa913649f8a1790800bf293694ed55b2ce8e322c81: Status 404 returned error can't find the container with id bfe1fbf708356b7449f08faa913649f8a1790800bf293694ed55b2ce8e322c81 Oct 02 10:01:21 crc kubenswrapper[4771]: I1002 10:01:21.632046 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-hh2km"] Oct 02 10:01:21 crc kubenswrapper[4771]: I1002 10:01:21.890862 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f7bdb69dd-8694c"] Oct 02 10:01:21 crc kubenswrapper[4771]: I1002 10:01:21.919794 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f7bdb69dd-8694c" event={"ID":"e17997e0-94d7-4a2f-83c1-dff9a23188cc","Type":"ContainerStarted","Data":"f85b46378cdbf0fa780c41165d4d193eaab1781fa5f3e2f5a351be23405e7c53"} Oct 02 10:01:21 crc kubenswrapper[4771]: I1002 10:01:21.922367 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" event={"ID":"cc8c97fe-a656-4e50-ae12-dd5a06a0a074","Type":"ContainerStarted","Data":"fc53ef2914755d376acec9eef516f290f24620ff4d895fcf0cc08dc0f5d4c52f"} Oct 02 10:01:21 crc kubenswrapper[4771]: I1002 10:01:21.924645 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64df995667-zkf22" event={"ID":"3439cba5-f86c-48f3-8d0b-587c02148975","Type":"ContainerStarted","Data":"cbd9bbebf8a984cd9c455234076171ce689b74bc82d9ff2ba93db8827bd19a96"} Oct 02 10:01:21 crc kubenswrapper[4771]: I1002 10:01:21.943343 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" event={"ID":"b69ba24b-3361-44d2-8774-64230ab7a376","Type":"ContainerStarted","Data":"bfe1fbf708356b7449f08faa913649f8a1790800bf293694ed55b2ce8e322c81"} Oct 02 10:01:22 crc kubenswrapper[4771]: I1002 10:01:22.994318 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f7bdb69dd-8694c" event={"ID":"e17997e0-94d7-4a2f-83c1-dff9a23188cc","Type":"ContainerStarted","Data":"f5cbb834ae1570d941b33c685c00f05ea5df87bc70a9d0e56f712cd404b3c40c"} Oct 02 10:01:22 crc kubenswrapper[4771]: I1002 10:01:22.994922 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f7bdb69dd-8694c" event={"ID":"e17997e0-94d7-4a2f-83c1-dff9a23188cc","Type":"ContainerStarted","Data":"83baa0af28908c2e64eec13434197211690a854fe2927d4cc4a5901c326094f2"} Oct 02 10:01:22 crc kubenswrapper[4771]: I1002 10:01:22.996278 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:22 crc kubenswrapper[4771]: I1002 10:01:22.996346 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:23 crc kubenswrapper[4771]: I1002 10:01:23.000518 4771 generic.go:334] "Generic (PLEG): container finished" podID="cc8c97fe-a656-4e50-ae12-dd5a06a0a074" containerID="1b30fe7ad8a8f83f928a5913bf04e38cb485c304d6c2bbe1ca10f5db0966fd14" exitCode=0 Oct 02 10:01:23 crc kubenswrapper[4771]: I1002 10:01:23.000578 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" event={"ID":"cc8c97fe-a656-4e50-ae12-dd5a06a0a074","Type":"ContainerDied","Data":"1b30fe7ad8a8f83f928a5913bf04e38cb485c304d6c2bbe1ca10f5db0966fd14"} Oct 02 10:01:23 crc kubenswrapper[4771]: I1002 10:01:23.048765 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5f7bdb69dd-8694c" podStartSLOduration=3.048738675 podStartE2EDuration="3.048738675s" podCreationTimestamp="2025-10-02 10:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:23.027719769 +0000 UTC m=+1470.675404846" watchObservedRunningTime="2025-10-02 10:01:23.048738675 +0000 UTC m=+1470.696423752" Oct 02 10:01:24 crc kubenswrapper[4771]: I1002 10:01:24.029417 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" event={"ID":"cc8c97fe-a656-4e50-ae12-dd5a06a0a074","Type":"ContainerStarted","Data":"91519164c8401612368d27576d9cc6221a369aafc1e574e609b90d954b89cf2c"} Oct 02 10:01:24 crc kubenswrapper[4771]: I1002 10:01:24.065428 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" podStartSLOduration=4.065400669 podStartE2EDuration="4.065400669s" podCreationTimestamp="2025-10-02 10:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:24.062947032 +0000 UTC m=+1471.710632099" watchObservedRunningTime="2025-10-02 10:01:24.065400669 +0000 UTC m=+1471.713085736" Oct 02 10:01:24 crc kubenswrapper[4771]: I1002 10:01:24.274937 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:24 crc kubenswrapper[4771]: I1002 10:01:24.375538 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:24 crc kubenswrapper[4771]: I1002 10:01:24.559945 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jmbf9"] Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.003382 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-c4746c7d4-mk657"] Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.010310 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.015891 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.016277 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.028733 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c4746c7d4-mk657"] Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.031774 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05fab4cb-323e-46a4-9b31-7e8f529e18cb-logs\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.031866 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-config-data-custom\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.031893 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-config-data\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.031951 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-public-tls-certs\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.032027 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-internal-tls-certs\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.032067 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-combined-ca-bundle\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.032194 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj72q\" (UniqueName: \"kubernetes.io/projected/05fab4cb-323e-46a4-9b31-7e8f529e18cb-kube-api-access-tj72q\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.084611 4771 generic.go:334] "Generic (PLEG): container finished" podID="2a276f13-9f48-40ae-8445-ce376fce62a1" containerID="f001212e1a5ae1a73fad0563d8f9c34c580f6d46c6321e4e686f1062692a8922" exitCode=0 Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.086348 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pkkwb" event={"ID":"2a276f13-9f48-40ae-8445-ce376fce62a1","Type":"ContainerDied","Data":"f001212e1a5ae1a73fad0563d8f9c34c580f6d46c6321e4e686f1062692a8922"} Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.088165 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.180598 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05fab4cb-323e-46a4-9b31-7e8f529e18cb-logs\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.181101 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-config-data-custom\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.181136 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-config-data\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.181259 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-public-tls-certs\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.183035 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05fab4cb-323e-46a4-9b31-7e8f529e18cb-logs\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.183196 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-internal-tls-certs\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.183550 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-combined-ca-bundle\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.183763 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj72q\" (UniqueName: \"kubernetes.io/projected/05fab4cb-323e-46a4-9b31-7e8f529e18cb-kube-api-access-tj72q\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.207893 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-config-data-custom\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.213283 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-combined-ca-bundle\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.214770 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-config-data\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.235009 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-public-tls-certs\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.238835 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05fab4cb-323e-46a4-9b31-7e8f529e18cb-internal-tls-certs\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.244668 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj72q\" (UniqueName: \"kubernetes.io/projected/05fab4cb-323e-46a4-9b31-7e8f529e18cb-kube-api-access-tj72q\") pod \"barbican-api-c4746c7d4-mk657\" (UID: \"05fab4cb-323e-46a4-9b31-7e8f529e18cb\") " pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:25 crc kubenswrapper[4771]: I1002 10:01:25.377445 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:26 crc kubenswrapper[4771]: I1002 10:01:26.096676 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jmbf9" podUID="47fce250-07d5-40e4-9a1d-009285bc460f" containerName="registry-server" containerID="cri-o://0c08b7daaa410e763a98ebf7c87dc9ea59f9163c7d40425a16028cecd7721502" gracePeriod=2 Oct 02 10:01:26 crc kubenswrapper[4771]: I1002 10:01:26.913692 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.124943 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-combined-ca-bundle\") pod \"2a276f13-9f48-40ae-8445-ce376fce62a1\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.125020 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a276f13-9f48-40ae-8445-ce376fce62a1-etc-machine-id\") pod \"2a276f13-9f48-40ae-8445-ce376fce62a1\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.125128 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-scripts\") pod \"2a276f13-9f48-40ae-8445-ce376fce62a1\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.125357 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-db-sync-config-data\") pod \"2a276f13-9f48-40ae-8445-ce376fce62a1\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.125481 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-config-data\") pod \"2a276f13-9f48-40ae-8445-ce376fce62a1\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.125570 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twr2n\" (UniqueName: \"kubernetes.io/projected/2a276f13-9f48-40ae-8445-ce376fce62a1-kube-api-access-twr2n\") pod \"2a276f13-9f48-40ae-8445-ce376fce62a1\" (UID: \"2a276f13-9f48-40ae-8445-ce376fce62a1\") " Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.128388 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a276f13-9f48-40ae-8445-ce376fce62a1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2a276f13-9f48-40ae-8445-ce376fce62a1" (UID: "2a276f13-9f48-40ae-8445-ce376fce62a1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.141874 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2a276f13-9f48-40ae-8445-ce376fce62a1" (UID: "2a276f13-9f48-40ae-8445-ce376fce62a1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.146040 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a276f13-9f48-40ae-8445-ce376fce62a1-kube-api-access-twr2n" (OuterVolumeSpecName: "kube-api-access-twr2n") pod "2a276f13-9f48-40ae-8445-ce376fce62a1" (UID: "2a276f13-9f48-40ae-8445-ce376fce62a1"). InnerVolumeSpecName "kube-api-access-twr2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.158219 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-scripts" (OuterVolumeSpecName: "scripts") pod "2a276f13-9f48-40ae-8445-ce376fce62a1" (UID: "2a276f13-9f48-40ae-8445-ce376fce62a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.170587 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.178881 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64df995667-zkf22" event={"ID":"3439cba5-f86c-48f3-8d0b-587c02148975","Type":"ContainerStarted","Data":"f244b7ee2ac400338b322ac56483a08aeb7bd03cb31ded20b072b71085dc6667"} Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.207662 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pkkwb" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.207723 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pkkwb" event={"ID":"2a276f13-9f48-40ae-8445-ce376fce62a1","Type":"ContainerDied","Data":"c7968dd590360b1c5058c83fad033a60ccc4433551e1bc26ee86520fb23e6fec"} Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.207763 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7968dd590360b1c5058c83fad033a60ccc4433551e1bc26ee86520fb23e6fec" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.244469 4771 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a276f13-9f48-40ae-8445-ce376fce62a1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.244510 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.244519 4771 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.244537 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twr2n\" (UniqueName: \"kubernetes.io/projected/2a276f13-9f48-40ae-8445-ce376fce62a1-kube-api-access-twr2n\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.282842 4771 generic.go:334] "Generic (PLEG): container finished" podID="47fce250-07d5-40e4-9a1d-009285bc460f" containerID="0c08b7daaa410e763a98ebf7c87dc9ea59f9163c7d40425a16028cecd7721502" exitCode=0 Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.282912 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jmbf9" event={"ID":"47fce250-07d5-40e4-9a1d-009285bc460f","Type":"ContainerDied","Data":"0c08b7daaa410e763a98ebf7c87dc9ea59f9163c7d40425a16028cecd7721502"} Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.282943 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jmbf9" event={"ID":"47fce250-07d5-40e4-9a1d-009285bc460f","Type":"ContainerDied","Data":"71eafb015a9e4a384843e8ffca77e869b4f0825753ca5c3deaf575bcace514b4"} Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.282978 4771 scope.go:117] "RemoveContainer" containerID="0c08b7daaa410e763a98ebf7c87dc9ea59f9163c7d40425a16028cecd7721502" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.283240 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jmbf9" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.307084 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-config-data" (OuterVolumeSpecName: "config-data") pod "2a276f13-9f48-40ae-8445-ce376fce62a1" (UID: "2a276f13-9f48-40ae-8445-ce376fce62a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.346608 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjf48\" (UniqueName: \"kubernetes.io/projected/47fce250-07d5-40e4-9a1d-009285bc460f-kube-api-access-hjf48\") pod \"47fce250-07d5-40e4-9a1d-009285bc460f\" (UID: \"47fce250-07d5-40e4-9a1d-009285bc460f\") " Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.347322 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47fce250-07d5-40e4-9a1d-009285bc460f-utilities\") pod \"47fce250-07d5-40e4-9a1d-009285bc460f\" (UID: \"47fce250-07d5-40e4-9a1d-009285bc460f\") " Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.347559 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47fce250-07d5-40e4-9a1d-009285bc460f-catalog-content\") pod \"47fce250-07d5-40e4-9a1d-009285bc460f\" (UID: \"47fce250-07d5-40e4-9a1d-009285bc460f\") " Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.350540 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47fce250-07d5-40e4-9a1d-009285bc460f-utilities" (OuterVolumeSpecName: "utilities") pod "47fce250-07d5-40e4-9a1d-009285bc460f" (UID: "47fce250-07d5-40e4-9a1d-009285bc460f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.354057 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.354116 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47fce250-07d5-40e4-9a1d-009285bc460f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.379500 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c4746c7d4-mk657"] Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.388578 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a276f13-9f48-40ae-8445-ce376fce62a1" (UID: "2a276f13-9f48-40ae-8445-ce376fce62a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.428454 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47fce250-07d5-40e4-9a1d-009285bc460f-kube-api-access-hjf48" (OuterVolumeSpecName: "kube-api-access-hjf48") pod "47fce250-07d5-40e4-9a1d-009285bc460f" (UID: "47fce250-07d5-40e4-9a1d-009285bc460f"). InnerVolumeSpecName "kube-api-access-hjf48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.457748 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a276f13-9f48-40ae-8445-ce376fce62a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.457790 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjf48\" (UniqueName: \"kubernetes.io/projected/47fce250-07d5-40e4-9a1d-009285bc460f-kube-api-access-hjf48\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.485267 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47fce250-07d5-40e4-9a1d-009285bc460f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47fce250-07d5-40e4-9a1d-009285bc460f" (UID: "47fce250-07d5-40e4-9a1d-009285bc460f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.519235 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:01:27 crc kubenswrapper[4771]: E1002 10:01:27.519882 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a276f13-9f48-40ae-8445-ce376fce62a1" containerName="cinder-db-sync" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.519900 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a276f13-9f48-40ae-8445-ce376fce62a1" containerName="cinder-db-sync" Oct 02 10:01:27 crc kubenswrapper[4771]: E1002 10:01:27.519917 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47fce250-07d5-40e4-9a1d-009285bc460f" containerName="extract-content" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.519924 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="47fce250-07d5-40e4-9a1d-009285bc460f" containerName="extract-content" Oct 02 10:01:27 crc kubenswrapper[4771]: E1002 10:01:27.519944 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47fce250-07d5-40e4-9a1d-009285bc460f" containerName="registry-server" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.519952 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="47fce250-07d5-40e4-9a1d-009285bc460f" containerName="registry-server" Oct 02 10:01:27 crc kubenswrapper[4771]: E1002 10:01:27.520008 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47fce250-07d5-40e4-9a1d-009285bc460f" containerName="extract-utilities" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.520016 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="47fce250-07d5-40e4-9a1d-009285bc460f" containerName="extract-utilities" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.520480 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a276f13-9f48-40ae-8445-ce376fce62a1" containerName="cinder-db-sync" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.520511 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="47fce250-07d5-40e4-9a1d-009285bc460f" containerName="registry-server" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.522089 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.528983 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.529290 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.529443 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jqk64" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.529627 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.559436 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.561375 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47fce250-07d5-40e4-9a1d-009285bc460f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.570683 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-hh2km"] Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.571277 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" podUID="cc8c97fe-a656-4e50-ae12-dd5a06a0a074" containerName="dnsmasq-dns" containerID="cri-o://91519164c8401612368d27576d9cc6221a369aafc1e574e609b90d954b89cf2c" gracePeriod=10 Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.670485 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.670553 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-scripts\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.670638 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-config-data\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.670843 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzsg4\" (UniqueName: \"kubernetes.io/projected/19df7b94-8a2c-43de-9a39-660d44eebe68-kube-api-access-nzsg4\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.670870 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/19df7b94-8a2c-43de-9a39-660d44eebe68-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.671200 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.732165 4771 scope.go:117] "RemoveContainer" containerID="07ad202028d929b0d5c7ae153b631470ea628deb7298ef99dbb197215eb227b6" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.782075 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.782604 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.782658 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-scripts\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.782732 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-config-data\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.782944 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzsg4\" (UniqueName: \"kubernetes.io/projected/19df7b94-8a2c-43de-9a39-660d44eebe68-kube-api-access-nzsg4\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.782983 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/19df7b94-8a2c-43de-9a39-660d44eebe68-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.791222 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/19df7b94-8a2c-43de-9a39-660d44eebe68-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.834098 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-scripts\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.860201 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-config-data\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.863810 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzsg4\" (UniqueName: \"kubernetes.io/projected/19df7b94-8a2c-43de-9a39-660d44eebe68-kube-api-access-nzsg4\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.865870 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-hjvjj"] Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.882813 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-hjvjj"] Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.882963 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.889162 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.890836 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:27 crc kubenswrapper[4771]: I1002 10:01:27.965938 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jmbf9"] Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.028442 4771 scope.go:117] "RemoveContainer" containerID="193d5055706cf3d956a16df74de9a60d2e1547f11d36babf2d507b3cde15055e" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.036717 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np2h9\" (UniqueName: \"kubernetes.io/projected/452a7728-1a14-4649-b5bd-2ea19e39db4c-kube-api-access-np2h9\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.036910 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-ovsdbserver-sb\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.036972 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-dns-swift-storage-0\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.037020 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-config\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.037055 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-dns-svc\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.037154 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-ovsdbserver-nb\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.091908 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jmbf9"] Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.144539 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np2h9\" (UniqueName: \"kubernetes.io/projected/452a7728-1a14-4649-b5bd-2ea19e39db4c-kube-api-access-np2h9\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.145011 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-ovsdbserver-sb\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.145072 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-dns-swift-storage-0\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.145114 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-config\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.145184 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-dns-svc\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.145270 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-ovsdbserver-nb\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.149624 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-ovsdbserver-nb\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.150932 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-ovsdbserver-sb\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.152379 4771 scope.go:117] "RemoveContainer" containerID="0c08b7daaa410e763a98ebf7c87dc9ea59f9163c7d40425a16028cecd7721502" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.153408 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-dns-swift-storage-0\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.153806 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-config\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: E1002 10:01:28.153912 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c08b7daaa410e763a98ebf7c87dc9ea59f9163c7d40425a16028cecd7721502\": container with ID starting with 0c08b7daaa410e763a98ebf7c87dc9ea59f9163c7d40425a16028cecd7721502 not found: ID does not exist" containerID="0c08b7daaa410e763a98ebf7c87dc9ea59f9163c7d40425a16028cecd7721502" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.153944 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c08b7daaa410e763a98ebf7c87dc9ea59f9163c7d40425a16028cecd7721502"} err="failed to get container status \"0c08b7daaa410e763a98ebf7c87dc9ea59f9163c7d40425a16028cecd7721502\": rpc error: code = NotFound desc = could not find container \"0c08b7daaa410e763a98ebf7c87dc9ea59f9163c7d40425a16028cecd7721502\": container with ID starting with 0c08b7daaa410e763a98ebf7c87dc9ea59f9163c7d40425a16028cecd7721502 not found: ID does not exist" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.153971 4771 scope.go:117] "RemoveContainer" containerID="07ad202028d929b0d5c7ae153b631470ea628deb7298ef99dbb197215eb227b6" Oct 02 10:01:28 crc kubenswrapper[4771]: E1002 10:01:28.155867 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07ad202028d929b0d5c7ae153b631470ea628deb7298ef99dbb197215eb227b6\": container with ID starting with 07ad202028d929b0d5c7ae153b631470ea628deb7298ef99dbb197215eb227b6 not found: ID does not exist" containerID="07ad202028d929b0d5c7ae153b631470ea628deb7298ef99dbb197215eb227b6" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.155903 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ad202028d929b0d5c7ae153b631470ea628deb7298ef99dbb197215eb227b6"} err="failed to get container status \"07ad202028d929b0d5c7ae153b631470ea628deb7298ef99dbb197215eb227b6\": rpc error: code = NotFound desc = could not find container \"07ad202028d929b0d5c7ae153b631470ea628deb7298ef99dbb197215eb227b6\": container with ID starting with 07ad202028d929b0d5c7ae153b631470ea628deb7298ef99dbb197215eb227b6 not found: ID does not exist" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.155927 4771 scope.go:117] "RemoveContainer" containerID="193d5055706cf3d956a16df74de9a60d2e1547f11d36babf2d507b3cde15055e" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.159939 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-dns-svc\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.170800 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.173698 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: E1002 10:01:28.176529 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"193d5055706cf3d956a16df74de9a60d2e1547f11d36babf2d507b3cde15055e\": container with ID starting with 193d5055706cf3d956a16df74de9a60d2e1547f11d36babf2d507b3cde15055e not found: ID does not exist" containerID="193d5055706cf3d956a16df74de9a60d2e1547f11d36babf2d507b3cde15055e" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.176578 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"193d5055706cf3d956a16df74de9a60d2e1547f11d36babf2d507b3cde15055e"} err="failed to get container status \"193d5055706cf3d956a16df74de9a60d2e1547f11d36babf2d507b3cde15055e\": rpc error: code = NotFound desc = could not find container \"193d5055706cf3d956a16df74de9a60d2e1547f11d36babf2d507b3cde15055e\": container with ID starting with 193d5055706cf3d956a16df74de9a60d2e1547f11d36babf2d507b3cde15055e not found: ID does not exist" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.184175 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.184921 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.186168 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.191506 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np2h9\" (UniqueName: \"kubernetes.io/projected/452a7728-1a14-4649-b5bd-2ea19e39db4c-kube-api-access-np2h9\") pod \"dnsmasq-dns-795f4db4bc-hjvjj\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.229667 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.251727 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c815084a-1f53-44b9-9e9b-97af81751781-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.251780 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-scripts\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.251832 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfccj\" (UniqueName: \"kubernetes.io/projected/c815084a-1f53-44b9-9e9b-97af81751781-kube-api-access-vfccj\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.251868 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c815084a-1f53-44b9-9e9b-97af81751781-logs\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.251899 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-config-data-custom\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.252051 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-config-data\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.252082 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.366613 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-config-data\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.366694 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.366956 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c815084a-1f53-44b9-9e9b-97af81751781-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.366984 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-scripts\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.367068 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfccj\" (UniqueName: \"kubernetes.io/projected/c815084a-1f53-44b9-9e9b-97af81751781-kube-api-access-vfccj\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.367114 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c815084a-1f53-44b9-9e9b-97af81751781-logs\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.380514 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c815084a-1f53-44b9-9e9b-97af81751781-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.381858 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-config-data-custom\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.415638 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c815084a-1f53-44b9-9e9b-97af81751781-logs\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.422473 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-config-data\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.422750 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.422962 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-scripts\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.433705 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfccj\" (UniqueName: \"kubernetes.io/projected/c815084a-1f53-44b9-9e9b-97af81751781-kube-api-access-vfccj\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.438888 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-config-data-custom\") pod \"cinder-api-0\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.452267 4771 generic.go:334] "Generic (PLEG): container finished" podID="cc8c97fe-a656-4e50-ae12-dd5a06a0a074" containerID="91519164c8401612368d27576d9cc6221a369aafc1e574e609b90d954b89cf2c" exitCode=0 Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.452376 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" event={"ID":"cc8c97fe-a656-4e50-ae12-dd5a06a0a074","Type":"ContainerDied","Data":"91519164c8401612368d27576d9cc6221a369aafc1e574e609b90d954b89cf2c"} Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.497435 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64df995667-zkf22" event={"ID":"3439cba5-f86c-48f3-8d0b-587c02148975","Type":"ContainerStarted","Data":"e4bb2e6431f11194c4c8decb10779ee37bc4fd20858c55d148c26b3339f85a87"} Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.549988 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.555671 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" event={"ID":"b69ba24b-3361-44d2-8774-64230ab7a376","Type":"ContainerStarted","Data":"b95f5bbc2bba1359bb749fefea0cede875b984082e53905f50162717d31072d8"} Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.576473 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c4746c7d4-mk657" event={"ID":"05fab4cb-323e-46a4-9b31-7e8f529e18cb","Type":"ContainerStarted","Data":"aac5015394fb3c0f648f182cf29d3ebf4d7ea5d57e43cd5acaa7eebf6987b14b"} Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.576542 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c4746c7d4-mk657" event={"ID":"05fab4cb-323e-46a4-9b31-7e8f529e18cb","Type":"ContainerStarted","Data":"1cc3616e4a79d1b53949dc1a175b2767e619ad60e7f96b301c978ae834cb63df"} Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.589292 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-64df995667-zkf22" podStartSLOduration=3.584832647 podStartE2EDuration="8.589263805s" podCreationTimestamp="2025-10-02 10:01:20 +0000 UTC" firstStartedPulling="2025-10-02 10:01:21.326471884 +0000 UTC m=+1468.974156951" lastFinishedPulling="2025-10-02 10:01:26.330903042 +0000 UTC m=+1473.978588109" observedRunningTime="2025-10-02 10:01:28.549598986 +0000 UTC m=+1476.197284053" watchObservedRunningTime="2025-10-02 10:01:28.589263805 +0000 UTC m=+1476.236948872" Oct 02 10:01:28 crc kubenswrapper[4771]: I1002 10:01:28.626516 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" podStartSLOduration=3.731103959 podStartE2EDuration="8.626486488s" podCreationTimestamp="2025-10-02 10:01:20 +0000 UTC" firstStartedPulling="2025-10-02 10:01:21.433270173 +0000 UTC m=+1469.080955230" lastFinishedPulling="2025-10-02 10:01:26.328652692 +0000 UTC m=+1473.976337759" observedRunningTime="2025-10-02 10:01:28.581953258 +0000 UTC m=+1476.229638325" watchObservedRunningTime="2025-10-02 10:01:28.626486488 +0000 UTC m=+1476.274171555" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.019465 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.108083 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-dns-swift-storage-0\") pod \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.108819 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-config\") pod \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.108957 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnztm\" (UniqueName: \"kubernetes.io/projected/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-kube-api-access-wnztm\") pod \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.109019 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-ovsdbserver-nb\") pod \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.109069 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-dns-svc\") pod \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.109105 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-ovsdbserver-sb\") pod \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\" (UID: \"cc8c97fe-a656-4e50-ae12-dd5a06a0a074\") " Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.176510 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-kube-api-access-wnztm" (OuterVolumeSpecName: "kube-api-access-wnztm") pod "cc8c97fe-a656-4e50-ae12-dd5a06a0a074" (UID: "cc8c97fe-a656-4e50-ae12-dd5a06a0a074"). InnerVolumeSpecName "kube-api-access-wnztm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.219030 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnztm\" (UniqueName: \"kubernetes.io/projected/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-kube-api-access-wnztm\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.514589 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cc8c97fe-a656-4e50-ae12-dd5a06a0a074" (UID: "cc8c97fe-a656-4e50-ae12-dd5a06a0a074"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.535769 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-config" (OuterVolumeSpecName: "config") pod "cc8c97fe-a656-4e50-ae12-dd5a06a0a074" (UID: "cc8c97fe-a656-4e50-ae12-dd5a06a0a074"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.538538 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cc8c97fe-a656-4e50-ae12-dd5a06a0a074" (UID: "cc8c97fe-a656-4e50-ae12-dd5a06a0a074"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.545353 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.545388 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.545397 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.546708 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cc8c97fe-a656-4e50-ae12-dd5a06a0a074" (UID: "cc8c97fe-a656-4e50-ae12-dd5a06a0a074"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.547243 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cc8c97fe-a656-4e50-ae12-dd5a06a0a074" (UID: "cc8c97fe-a656-4e50-ae12-dd5a06a0a074"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.640735 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" event={"ID":"cc8c97fe-a656-4e50-ae12-dd5a06a0a074","Type":"ContainerDied","Data":"fc53ef2914755d376acec9eef516f290f24620ff4d895fcf0cc08dc0f5d4c52f"} Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.640832 4771 scope.go:117] "RemoveContainer" containerID="91519164c8401612368d27576d9cc6221a369aafc1e574e609b90d954b89cf2c" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.641029 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-hh2km" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.648048 4771 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.648093 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc8c97fe-a656-4e50-ae12-dd5a06a0a074-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.675263 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6998cd89db-l2xv7" event={"ID":"b69ba24b-3361-44d2-8774-64230ab7a376","Type":"ContainerStarted","Data":"92385d1dac7b71554cf842aa26f0ad74addcd5c207c64154a9cd9109dd82b46c"} Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.851122 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47fce250-07d5-40e4-9a1d-009285bc460f" path="/var/lib/kubelet/pods/47fce250-07d5-40e4-9a1d-009285bc460f/volumes" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.853186 4771 scope.go:117] "RemoveContainer" containerID="1b30fe7ad8a8f83f928a5913bf04e38cb485c304d6c2bbe1ca10f5db0966fd14" Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.857832 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-hjvjj"] Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.857882 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.857899 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-hh2km"] Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.857917 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-hh2km"] Oct 02 10:01:29 crc kubenswrapper[4771]: W1002 10:01:29.875532 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod452a7728_1a14_4649_b5bd_2ea19e39db4c.slice/crio-64d899e416a2b671f0fa00b71531d345ac16cc99372eb45159906bddce6247b0 WatchSource:0}: Error finding container 64d899e416a2b671f0fa00b71531d345ac16cc99372eb45159906bddce6247b0: Status 404 returned error can't find the container with id 64d899e416a2b671f0fa00b71531d345ac16cc99372eb45159906bddce6247b0 Oct 02 10:01:29 crc kubenswrapper[4771]: I1002 10:01:29.988505 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:01:30 crc kubenswrapper[4771]: I1002 10:01:30.726177 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"19df7b94-8a2c-43de-9a39-660d44eebe68","Type":"ContainerStarted","Data":"1df272aa123244b7a9fff24e9c68204f17576c71543391f6197dfa6cd9b03b0f"} Oct 02 10:01:30 crc kubenswrapper[4771]: I1002 10:01:30.741795 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c4746c7d4-mk657" event={"ID":"05fab4cb-323e-46a4-9b31-7e8f529e18cb","Type":"ContainerStarted","Data":"cd46e54a1976621b08ecc4237dcfac29f51d1c98fc34e870624abaf8563a3a64"} Oct 02 10:01:30 crc kubenswrapper[4771]: I1002 10:01:30.742831 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:30 crc kubenswrapper[4771]: I1002 10:01:30.742918 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:30 crc kubenswrapper[4771]: I1002 10:01:30.754254 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c815084a-1f53-44b9-9e9b-97af81751781","Type":"ContainerStarted","Data":"b3670a8abe27df3f8100d22987541594b3e4ad18477e5f4c2c5dfdfac7f79c19"} Oct 02 10:01:30 crc kubenswrapper[4771]: I1002 10:01:30.765527 4771 generic.go:334] "Generic (PLEG): container finished" podID="452a7728-1a14-4649-b5bd-2ea19e39db4c" containerID="ea940924145b3c113e05d565e96d23b8813882966a0caa89309309a12e344864" exitCode=0 Oct 02 10:01:30 crc kubenswrapper[4771]: I1002 10:01:30.765923 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" event={"ID":"452a7728-1a14-4649-b5bd-2ea19e39db4c","Type":"ContainerDied","Data":"ea940924145b3c113e05d565e96d23b8813882966a0caa89309309a12e344864"} Oct 02 10:01:30 crc kubenswrapper[4771]: I1002 10:01:30.765985 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" event={"ID":"452a7728-1a14-4649-b5bd-2ea19e39db4c","Type":"ContainerStarted","Data":"64d899e416a2b671f0fa00b71531d345ac16cc99372eb45159906bddce6247b0"} Oct 02 10:01:30 crc kubenswrapper[4771]: I1002 10:01:30.788499 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-c4746c7d4-mk657" podStartSLOduration=6.788472222 podStartE2EDuration="6.788472222s" podCreationTimestamp="2025-10-02 10:01:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:30.778896054 +0000 UTC m=+1478.426581121" watchObservedRunningTime="2025-10-02 10:01:30.788472222 +0000 UTC m=+1478.436157299" Oct 02 10:01:31 crc kubenswrapper[4771]: I1002 10:01:31.173426 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j4c5j" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="registry-server" probeResult="failure" output=< Oct 02 10:01:31 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:01:31 crc kubenswrapper[4771]: > Oct 02 10:01:31 crc kubenswrapper[4771]: I1002 10:01:31.738726 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc8c97fe-a656-4e50-ae12-dd5a06a0a074" path="/var/lib/kubelet/pods/cc8c97fe-a656-4e50-ae12-dd5a06a0a074/volumes" Oct 02 10:01:31 crc kubenswrapper[4771]: I1002 10:01:31.930431 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c815084a-1f53-44b9-9e9b-97af81751781","Type":"ContainerStarted","Data":"83e78f7da627477e1552fd241e153c4f62e3378550cf4e8c1e8c849e67fbba56"} Oct 02 10:01:31 crc kubenswrapper[4771]: I1002 10:01:31.962181 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" event={"ID":"452a7728-1a14-4649-b5bd-2ea19e39db4c","Type":"ContainerStarted","Data":"9be0f1e2a3ff9da209f926f223a0fcd17605dd544224d89e0909b5bb670ab455"} Oct 02 10:01:31 crc kubenswrapper[4771]: I1002 10:01:31.962261 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:32 crc kubenswrapper[4771]: I1002 10:01:32.017112 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" podStartSLOduration=5.017088629 podStartE2EDuration="5.017088629s" podCreationTimestamp="2025-10-02 10:01:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:32.012074484 +0000 UTC m=+1479.659759561" watchObservedRunningTime="2025-10-02 10:01:32.017088629 +0000 UTC m=+1479.664773706" Oct 02 10:01:33 crc kubenswrapper[4771]: I1002 10:01:33.020062 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:01:33 crc kubenswrapper[4771]: I1002 10:01:33.295958 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-758b4b5558-442kz" Oct 02 10:01:34 crc kubenswrapper[4771]: I1002 10:01:34.051514 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c815084a-1f53-44b9-9e9b-97af81751781","Type":"ContainerStarted","Data":"e43489073b64230543bc59159b4ba0a8a399eb2c467524bc0c048d6b7f059114"} Oct 02 10:01:34 crc kubenswrapper[4771]: I1002 10:01:34.052468 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="c815084a-1f53-44b9-9e9b-97af81751781" containerName="cinder-api-log" containerID="cri-o://83e78f7da627477e1552fd241e153c4f62e3378550cf4e8c1e8c849e67fbba56" gracePeriod=30 Oct 02 10:01:34 crc kubenswrapper[4771]: I1002 10:01:34.052593 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 10:01:34 crc kubenswrapper[4771]: I1002 10:01:34.053320 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="c815084a-1f53-44b9-9e9b-97af81751781" containerName="cinder-api" containerID="cri-o://e43489073b64230543bc59159b4ba0a8a399eb2c467524bc0c048d6b7f059114" gracePeriod=30 Oct 02 10:01:34 crc kubenswrapper[4771]: I1002 10:01:34.063815 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"19df7b94-8a2c-43de-9a39-660d44eebe68","Type":"ContainerStarted","Data":"2a6dc00621bb08127ae6b9ba1f1abe66022c0921e7a8d72e8c2d651ec7d578ee"} Oct 02 10:01:34 crc kubenswrapper[4771]: I1002 10:01:34.098192 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.098166642 podStartE2EDuration="7.098166642s" podCreationTimestamp="2025-10-02 10:01:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:34.080263889 +0000 UTC m=+1481.727948976" watchObservedRunningTime="2025-10-02 10:01:34.098166642 +0000 UTC m=+1481.745851709" Oct 02 10:01:34 crc kubenswrapper[4771]: I1002 10:01:34.989607 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5f7bdb69dd-8694c" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.201:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 10:01:35 crc kubenswrapper[4771]: I1002 10:01:35.130877 4771 generic.go:334] "Generic (PLEG): container finished" podID="c815084a-1f53-44b9-9e9b-97af81751781" containerID="83e78f7da627477e1552fd241e153c4f62e3378550cf4e8c1e8c849e67fbba56" exitCode=143 Oct 02 10:01:35 crc kubenswrapper[4771]: I1002 10:01:35.130942 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c815084a-1f53-44b9-9e9b-97af81751781","Type":"ContainerDied","Data":"83e78f7da627477e1552fd241e153c4f62e3378550cf4e8c1e8c849e67fbba56"} Oct 02 10:01:35 crc kubenswrapper[4771]: I1002 10:01:35.284607 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:35 crc kubenswrapper[4771]: I1002 10:01:35.887386 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:01:35 crc kubenswrapper[4771]: I1002 10:01:35.974021 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f7bdb69dd-8694c" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.201:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.040091 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c815084a-1f53-44b9-9e9b-97af81751781-etc-machine-id\") pod \"c815084a-1f53-44b9-9e9b-97af81751781\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.040233 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-config-data-custom\") pod \"c815084a-1f53-44b9-9e9b-97af81751781\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.040770 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-scripts\") pod \"c815084a-1f53-44b9-9e9b-97af81751781\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.040873 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-combined-ca-bundle\") pod \"c815084a-1f53-44b9-9e9b-97af81751781\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.040939 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c815084a-1f53-44b9-9e9b-97af81751781-logs\") pod \"c815084a-1f53-44b9-9e9b-97af81751781\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.041087 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfccj\" (UniqueName: \"kubernetes.io/projected/c815084a-1f53-44b9-9e9b-97af81751781-kube-api-access-vfccj\") pod \"c815084a-1f53-44b9-9e9b-97af81751781\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.041197 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-config-data\") pod \"c815084a-1f53-44b9-9e9b-97af81751781\" (UID: \"c815084a-1f53-44b9-9e9b-97af81751781\") " Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.042252 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c815084a-1f53-44b9-9e9b-97af81751781-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c815084a-1f53-44b9-9e9b-97af81751781" (UID: "c815084a-1f53-44b9-9e9b-97af81751781"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.042571 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c815084a-1f53-44b9-9e9b-97af81751781-logs" (OuterVolumeSpecName: "logs") pod "c815084a-1f53-44b9-9e9b-97af81751781" (UID: "c815084a-1f53-44b9-9e9b-97af81751781"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.072957 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c815084a-1f53-44b9-9e9b-97af81751781-kube-api-access-vfccj" (OuterVolumeSpecName: "kube-api-access-vfccj") pod "c815084a-1f53-44b9-9e9b-97af81751781" (UID: "c815084a-1f53-44b9-9e9b-97af81751781"). InnerVolumeSpecName "kube-api-access-vfccj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.072967 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-scripts" (OuterVolumeSpecName: "scripts") pod "c815084a-1f53-44b9-9e9b-97af81751781" (UID: "c815084a-1f53-44b9-9e9b-97af81751781"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.088454 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c815084a-1f53-44b9-9e9b-97af81751781" (UID: "c815084a-1f53-44b9-9e9b-97af81751781"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.135288 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c815084a-1f53-44b9-9e9b-97af81751781" (UID: "c815084a-1f53-44b9-9e9b-97af81751781"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.150105 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfccj\" (UniqueName: \"kubernetes.io/projected/c815084a-1f53-44b9-9e9b-97af81751781-kube-api-access-vfccj\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.150188 4771 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c815084a-1f53-44b9-9e9b-97af81751781-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.150203 4771 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.150217 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.150230 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.150240 4771 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c815084a-1f53-44b9-9e9b-97af81751781-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.166899 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"19df7b94-8a2c-43de-9a39-660d44eebe68","Type":"ContainerStarted","Data":"5c4bb2cb61bb8786404036fbd098013c3755be286080fc8ef6bf08caa9c69e4a"} Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.175332 4771 generic.go:334] "Generic (PLEG): container finished" podID="c815084a-1f53-44b9-9e9b-97af81751781" containerID="e43489073b64230543bc59159b4ba0a8a399eb2c467524bc0c048d6b7f059114" exitCode=0 Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.175396 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c815084a-1f53-44b9-9e9b-97af81751781","Type":"ContainerDied","Data":"e43489073b64230543bc59159b4ba0a8a399eb2c467524bc0c048d6b7f059114"} Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.175429 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c815084a-1f53-44b9-9e9b-97af81751781","Type":"ContainerDied","Data":"b3670a8abe27df3f8100d22987541594b3e4ad18477e5f4c2c5dfdfac7f79c19"} Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.175458 4771 scope.go:117] "RemoveContainer" containerID="e43489073b64230543bc59159b4ba0a8a399eb2c467524bc0c048d6b7f059114" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.175640 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.199012 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-config-data" (OuterVolumeSpecName: "config-data") pod "c815084a-1f53-44b9-9e9b-97af81751781" (UID: "c815084a-1f53-44b9-9e9b-97af81751781"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.202429 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.76464995 podStartE2EDuration="9.202407859s" podCreationTimestamp="2025-10-02 10:01:27 +0000 UTC" firstStartedPulling="2025-10-02 10:01:29.990904664 +0000 UTC m=+1477.638589721" lastFinishedPulling="2025-10-02 10:01:32.428662563 +0000 UTC m=+1480.076347630" observedRunningTime="2025-10-02 10:01:36.192404569 +0000 UTC m=+1483.840089636" watchObservedRunningTime="2025-10-02 10:01:36.202407859 +0000 UTC m=+1483.850092926" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.252466 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c815084a-1f53-44b9-9e9b-97af81751781-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.327334 4771 scope.go:117] "RemoveContainer" containerID="83e78f7da627477e1552fd241e153c4f62e3378550cf4e8c1e8c849e67fbba56" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.377769 4771 scope.go:117] "RemoveContainer" containerID="e43489073b64230543bc59159b4ba0a8a399eb2c467524bc0c048d6b7f059114" Oct 02 10:01:36 crc kubenswrapper[4771]: E1002 10:01:36.378551 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e43489073b64230543bc59159b4ba0a8a399eb2c467524bc0c048d6b7f059114\": container with ID starting with e43489073b64230543bc59159b4ba0a8a399eb2c467524bc0c048d6b7f059114 not found: ID does not exist" containerID="e43489073b64230543bc59159b4ba0a8a399eb2c467524bc0c048d6b7f059114" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.378600 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e43489073b64230543bc59159b4ba0a8a399eb2c467524bc0c048d6b7f059114"} err="failed to get container status \"e43489073b64230543bc59159b4ba0a8a399eb2c467524bc0c048d6b7f059114\": rpc error: code = NotFound desc = could not find container \"e43489073b64230543bc59159b4ba0a8a399eb2c467524bc0c048d6b7f059114\": container with ID starting with e43489073b64230543bc59159b4ba0a8a399eb2c467524bc0c048d6b7f059114 not found: ID does not exist" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.378632 4771 scope.go:117] "RemoveContainer" containerID="83e78f7da627477e1552fd241e153c4f62e3378550cf4e8c1e8c849e67fbba56" Oct 02 10:01:36 crc kubenswrapper[4771]: E1002 10:01:36.379300 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83e78f7da627477e1552fd241e153c4f62e3378550cf4e8c1e8c849e67fbba56\": container with ID starting with 83e78f7da627477e1552fd241e153c4f62e3378550cf4e8c1e8c849e67fbba56 not found: ID does not exist" containerID="83e78f7da627477e1552fd241e153c4f62e3378550cf4e8c1e8c849e67fbba56" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.379369 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83e78f7da627477e1552fd241e153c4f62e3378550cf4e8c1e8c849e67fbba56"} err="failed to get container status \"83e78f7da627477e1552fd241e153c4f62e3378550cf4e8c1e8c849e67fbba56\": rpc error: code = NotFound desc = could not find container \"83e78f7da627477e1552fd241e153c4f62e3378550cf4e8c1e8c849e67fbba56\": container with ID starting with 83e78f7da627477e1552fd241e153c4f62e3378550cf4e8c1e8c849e67fbba56 not found: ID does not exist" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.528009 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.540440 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.546832 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.565225 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:01:36 crc kubenswrapper[4771]: E1002 10:01:36.566016 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8c97fe-a656-4e50-ae12-dd5a06a0a074" containerName="init" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.566061 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8c97fe-a656-4e50-ae12-dd5a06a0a074" containerName="init" Oct 02 10:01:36 crc kubenswrapper[4771]: E1002 10:01:36.566083 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c815084a-1f53-44b9-9e9b-97af81751781" containerName="cinder-api-log" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.566092 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="c815084a-1f53-44b9-9e9b-97af81751781" containerName="cinder-api-log" Oct 02 10:01:36 crc kubenswrapper[4771]: E1002 10:01:36.566114 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c815084a-1f53-44b9-9e9b-97af81751781" containerName="cinder-api" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.566126 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="c815084a-1f53-44b9-9e9b-97af81751781" containerName="cinder-api" Oct 02 10:01:36 crc kubenswrapper[4771]: E1002 10:01:36.566194 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8c97fe-a656-4e50-ae12-dd5a06a0a074" containerName="dnsmasq-dns" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.566205 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8c97fe-a656-4e50-ae12-dd5a06a0a074" containerName="dnsmasq-dns" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.566499 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="c815084a-1f53-44b9-9e9b-97af81751781" containerName="cinder-api-log" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.566538 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="c815084a-1f53-44b9-9e9b-97af81751781" containerName="cinder-api" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.566563 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc8c97fe-a656-4e50-ae12-dd5a06a0a074" containerName="dnsmasq-dns" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.568378 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.573719 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.574045 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.574233 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.580693 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.694732 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.694896 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-logs\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.694954 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkmn4\" (UniqueName: \"kubernetes.io/projected/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-kube-api-access-zkmn4\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.695020 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-scripts\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.695059 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.695110 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-public-tls-certs\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.698623 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.698787 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-config-data-custom\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.698885 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-config-data\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.801807 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.801916 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-logs\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.801975 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkmn4\" (UniqueName: \"kubernetes.io/projected/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-kube-api-access-zkmn4\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.802013 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-scripts\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.802039 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.802074 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-public-tls-certs\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.802091 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.802122 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-config-data-custom\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.802260 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-config-data\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.802581 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.803367 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-logs\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.811504 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-scripts\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.812171 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.814676 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-public-tls-certs\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.818814 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.822367 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-config-data-custom\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.829698 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkmn4\" (UniqueName: \"kubernetes.io/projected/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-kube-api-access-zkmn4\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.832366 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b3670c-72d0-4a3d-aa68-f0b69f9653e9-config-data\") pod \"cinder-api-0\" (UID: \"55b3670c-72d0-4a3d-aa68-f0b69f9653e9\") " pod="openstack/cinder-api-0" Oct 02 10:01:36 crc kubenswrapper[4771]: I1002 10:01:36.915789 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.244434 4771 generic.go:334] "Generic (PLEG): container finished" podID="975630d7-f63d-43eb-99c4-a245f4f20c24" containerID="0bf3f374add61e91e09dea2502ed16444da7aec47589d12eb1132079e99b1dec" exitCode=0 Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.246242 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-f725t" event={"ID":"975630d7-f63d-43eb-99c4-a245f4f20c24","Type":"ContainerDied","Data":"0bf3f374add61e91e09dea2502ed16444da7aec47589d12eb1132079e99b1dec"} Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.616418 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.706619 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c815084a-1f53-44b9-9e9b-97af81751781" path="/var/lib/kubelet/pods/c815084a-1f53-44b9-9e9b-97af81751781/volumes" Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.763932 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-svb6n"] Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.767667 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.785534 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svb6n"] Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.835278 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-catalog-content\") pod \"community-operators-svb6n\" (UID: \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\") " pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.835354 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zn4t\" (UniqueName: \"kubernetes.io/projected/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-kube-api-access-9zn4t\") pod \"community-operators-svb6n\" (UID: \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\") " pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.835650 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-utilities\") pod \"community-operators-svb6n\" (UID: \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\") " pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.937067 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-catalog-content\") pod \"community-operators-svb6n\" (UID: \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\") " pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.937152 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zn4t\" (UniqueName: \"kubernetes.io/projected/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-kube-api-access-9zn4t\") pod \"community-operators-svb6n\" (UID: \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\") " pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.937270 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-utilities\") pod \"community-operators-svb6n\" (UID: \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\") " pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.937860 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-catalog-content\") pod \"community-operators-svb6n\" (UID: \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\") " pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.937958 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-utilities\") pod \"community-operators-svb6n\" (UID: \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\") " pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:37 crc kubenswrapper[4771]: I1002 10:01:37.972038 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zn4t\" (UniqueName: \"kubernetes.io/projected/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-kube-api-access-9zn4t\") pod \"community-operators-svb6n\" (UID: \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\") " pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.097181 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.098950 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.102175 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.102953 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-zzhd7" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.103154 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.110579 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.128812 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.185975 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.232795 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.246078 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91134808-2774-4b54-8a59-09e9447dd87f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"91134808-2774-4b54-8a59-09e9447dd87f\") " pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.246305 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/91134808-2774-4b54-8a59-09e9447dd87f-openstack-config\") pod \"openstackclient\" (UID: \"91134808-2774-4b54-8a59-09e9447dd87f\") " pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.246358 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxxsd\" (UniqueName: \"kubernetes.io/projected/91134808-2774-4b54-8a59-09e9447dd87f-kube-api-access-lxxsd\") pod \"openstackclient\" (UID: \"91134808-2774-4b54-8a59-09e9447dd87f\") " pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.246392 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/91134808-2774-4b54-8a59-09e9447dd87f-openstack-config-secret\") pod \"openstackclient\" (UID: \"91134808-2774-4b54-8a59-09e9447dd87f\") " pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.297876 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"55b3670c-72d0-4a3d-aa68-f0b69f9653e9","Type":"ContainerStarted","Data":"ca94cecce6338520dc67f3402d4c90f3ef06a2cd50bfefb02a6dc3d6f0ad93e8"} Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.332746 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-qr4df"] Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.337447 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" podUID="eb11e684-5673-479e-873b-d61930fe2d52" containerName="dnsmasq-dns" containerID="cri-o://31ea447f7214766a6d3dd9b45518a5143478760e9e91f9ee76e181ab8b6af100" gracePeriod=10 Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.354053 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/91134808-2774-4b54-8a59-09e9447dd87f-openstack-config\") pod \"openstackclient\" (UID: \"91134808-2774-4b54-8a59-09e9447dd87f\") " pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.354699 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxxsd\" (UniqueName: \"kubernetes.io/projected/91134808-2774-4b54-8a59-09e9447dd87f-kube-api-access-lxxsd\") pod \"openstackclient\" (UID: \"91134808-2774-4b54-8a59-09e9447dd87f\") " pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.354866 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/91134808-2774-4b54-8a59-09e9447dd87f-openstack-config-secret\") pod \"openstackclient\" (UID: \"91134808-2774-4b54-8a59-09e9447dd87f\") " pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.354998 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91134808-2774-4b54-8a59-09e9447dd87f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"91134808-2774-4b54-8a59-09e9447dd87f\") " pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.355960 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/91134808-2774-4b54-8a59-09e9447dd87f-openstack-config\") pod \"openstackclient\" (UID: \"91134808-2774-4b54-8a59-09e9447dd87f\") " pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.379554 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/91134808-2774-4b54-8a59-09e9447dd87f-openstack-config-secret\") pod \"openstackclient\" (UID: \"91134808-2774-4b54-8a59-09e9447dd87f\") " pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.380426 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91134808-2774-4b54-8a59-09e9447dd87f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"91134808-2774-4b54-8a59-09e9447dd87f\") " pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.389033 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxxsd\" (UniqueName: \"kubernetes.io/projected/91134808-2774-4b54-8a59-09e9447dd87f-kube-api-access-lxxsd\") pod \"openstackclient\" (UID: \"91134808-2774-4b54-8a59-09e9447dd87f\") " pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.447501 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.683147 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:38 crc kubenswrapper[4771]: I1002 10:01:38.827705 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" podUID="eb11e684-5673-479e-873b-d61930fe2d52" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.186:5353: connect: connection refused" Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.048732 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.283437 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-f725t" Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.324751 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vckbr\" (UniqueName: \"kubernetes.io/projected/975630d7-f63d-43eb-99c4-a245f4f20c24-kube-api-access-vckbr\") pod \"975630d7-f63d-43eb-99c4-a245f4f20c24\" (UID: \"975630d7-f63d-43eb-99c4-a245f4f20c24\") " Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.329778 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975630d7-f63d-43eb-99c4-a245f4f20c24-combined-ca-bundle\") pod \"975630d7-f63d-43eb-99c4-a245f4f20c24\" (UID: \"975630d7-f63d-43eb-99c4-a245f4f20c24\") " Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.331909 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/975630d7-f63d-43eb-99c4-a245f4f20c24-config\") pod \"975630d7-f63d-43eb-99c4-a245f4f20c24\" (UID: \"975630d7-f63d-43eb-99c4-a245f4f20c24\") " Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.342854 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/975630d7-f63d-43eb-99c4-a245f4f20c24-kube-api-access-vckbr" (OuterVolumeSpecName: "kube-api-access-vckbr") pod "975630d7-f63d-43eb-99c4-a245f4f20c24" (UID: "975630d7-f63d-43eb-99c4-a245f4f20c24"). InnerVolumeSpecName "kube-api-access-vckbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.412824 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-f725t" event={"ID":"975630d7-f63d-43eb-99c4-a245f4f20c24","Type":"ContainerDied","Data":"49cb0c97d1134b45170e0b252cc621b8723ed6162dd48c90c356e511fd4df31e"} Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.415009 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49cb0c97d1134b45170e0b252cc621b8723ed6162dd48c90c356e511fd4df31e" Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.412905 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-f725t" Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.449680 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975630d7-f63d-43eb-99c4-a245f4f20c24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "975630d7-f63d-43eb-99c4-a245f4f20c24" (UID: "975630d7-f63d-43eb-99c4-a245f4f20c24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.455709 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"55b3670c-72d0-4a3d-aa68-f0b69f9653e9","Type":"ContainerStarted","Data":"8d7984080b76bba9594ccb3f014f216da952c920e2d3a583680767531b1f80a5"} Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.470439 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vckbr\" (UniqueName: \"kubernetes.io/projected/975630d7-f63d-43eb-99c4-a245f4f20c24-kube-api-access-vckbr\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.470502 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975630d7-f63d-43eb-99c4-a245f4f20c24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.475376 4771 generic.go:334] "Generic (PLEG): container finished" podID="eb11e684-5673-479e-873b-d61930fe2d52" containerID="31ea447f7214766a6d3dd9b45518a5143478760e9e91f9ee76e181ab8b6af100" exitCode=0 Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.476093 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" event={"ID":"eb11e684-5673-479e-873b-d61930fe2d52","Type":"ContainerDied","Data":"31ea447f7214766a6d3dd9b45518a5143478760e9e91f9ee76e181ab8b6af100"} Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.481700 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975630d7-f63d-43eb-99c4-a245f4f20c24-config" (OuterVolumeSpecName: "config") pod "975630d7-f63d-43eb-99c4-a245f4f20c24" (UID: "975630d7-f63d-43eb-99c4-a245f4f20c24"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.571110 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.573937 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/975630d7-f63d-43eb-99c4-a245f4f20c24-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.673015 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c4746c7d4-mk657" Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.845283 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f7bdb69dd-8694c"] Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.845577 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f7bdb69dd-8694c" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerName="barbican-api-log" containerID="cri-o://83baa0af28908c2e64eec13434197211690a854fe2927d4cc4a5901c326094f2" gracePeriod=30 Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.845786 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f7bdb69dd-8694c" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerName="barbican-api" containerID="cri-o://f5cbb834ae1570d941b33c685c00f05ea5df87bc70a9d0e56f712cd404b3c40c" gracePeriod=30 Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.862700 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f7bdb69dd-8694c" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.201:9311/healthcheck\": EOF" Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.913354 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 10:01:39 crc kubenswrapper[4771]: I1002 10:01:39.982266 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svb6n"] Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.406214 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.515848 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"91134808-2774-4b54-8a59-09e9447dd87f","Type":"ContainerStarted","Data":"b6ee5a10204c8c35b7ab2eed45af88dd367bba3482108733a444be44867e34f1"} Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.523609 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-nb\") pod \"eb11e684-5673-479e-873b-d61930fe2d52\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.523699 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-config\") pod \"eb11e684-5673-479e-873b-d61930fe2d52\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.523835 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf9xw\" (UniqueName: \"kubernetes.io/projected/eb11e684-5673-479e-873b-d61930fe2d52-kube-api-access-kf9xw\") pod \"eb11e684-5673-479e-873b-d61930fe2d52\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.523864 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-dns-swift-storage-0\") pod \"eb11e684-5673-479e-873b-d61930fe2d52\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.523999 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-dns-svc\") pod \"eb11e684-5673-479e-873b-d61930fe2d52\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.524085 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-sb\") pod \"eb11e684-5673-479e-873b-d61930fe2d52\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.562867 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" event={"ID":"eb11e684-5673-479e-873b-d61930fe2d52","Type":"ContainerDied","Data":"d3ca1d08001b7497684e89f49b7e32471cca11c470a5dd03a11c310b44e9e9e5"} Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.563214 4771 scope.go:117] "RemoveContainer" containerID="31ea447f7214766a6d3dd9b45518a5143478760e9e91f9ee76e181ab8b6af100" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.563528 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-qr4df" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.598242 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-chfb4"] Oct 02 10:01:40 crc kubenswrapper[4771]: E1002 10:01:40.598754 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="975630d7-f63d-43eb-99c4-a245f4f20c24" containerName="neutron-db-sync" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.598771 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="975630d7-f63d-43eb-99c4-a245f4f20c24" containerName="neutron-db-sync" Oct 02 10:01:40 crc kubenswrapper[4771]: E1002 10:01:40.598785 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb11e684-5673-479e-873b-d61930fe2d52" containerName="dnsmasq-dns" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.598794 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb11e684-5673-479e-873b-d61930fe2d52" containerName="dnsmasq-dns" Oct 02 10:01:40 crc kubenswrapper[4771]: E1002 10:01:40.598840 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb11e684-5673-479e-873b-d61930fe2d52" containerName="init" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.598847 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb11e684-5673-479e-873b-d61930fe2d52" containerName="init" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.599057 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb11e684-5673-479e-873b-d61930fe2d52" containerName="dnsmasq-dns" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.599092 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="975630d7-f63d-43eb-99c4-a245f4f20c24" containerName="neutron-db-sync" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.616447 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.643403 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb11e684-5673-479e-873b-d61930fe2d52-kube-api-access-kf9xw" (OuterVolumeSpecName: "kube-api-access-kf9xw") pod "eb11e684-5673-479e-873b-d61930fe2d52" (UID: "eb11e684-5673-479e-873b-d61930fe2d52"). InnerVolumeSpecName "kube-api-access-kf9xw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.646512 4771 generic.go:334] "Generic (PLEG): container finished" podID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerID="83baa0af28908c2e64eec13434197211690a854fe2927d4cc4a5901c326094f2" exitCode=143 Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.646687 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f7bdb69dd-8694c" event={"ID":"e17997e0-94d7-4a2f-83c1-dff9a23188cc","Type":"ContainerDied","Data":"83baa0af28908c2e64eec13434197211690a854fe2927d4cc4a5901c326094f2"} Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.650470 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf9xw\" (UniqueName: \"kubernetes.io/projected/eb11e684-5673-479e-873b-d61930fe2d52-kube-api-access-kf9xw\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.665885 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="19df7b94-8a2c-43de-9a39-660d44eebe68" containerName="cinder-scheduler" containerID="cri-o://2a6dc00621bb08127ae6b9ba1f1abe66022c0921e7a8d72e8c2d651ec7d578ee" gracePeriod=30 Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.666050 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svb6n" event={"ID":"ec78b084-92a4-4eb9-ac22-bd7b51f4612c","Type":"ContainerStarted","Data":"976fba3bb30662f94c2388470dbee4ab0496bf29e0dbf1a6b38e0be5299e6841"} Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.666116 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="19df7b94-8a2c-43de-9a39-660d44eebe68" containerName="probe" containerID="cri-o://5c4bb2cb61bb8786404036fbd098013c3755be286080fc8ef6bf08caa9c69e4a" gracePeriod=30 Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.758064 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-chfb4"] Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.767498 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-config\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.767638 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wtgm\" (UniqueName: \"kubernetes.io/projected/7240f9bd-432f-492f-b83c-d1e265f34505-kube-api-access-4wtgm\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.767730 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.767962 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.769340 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.769701 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.814658 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7d8b77dfb-5n9ct"] Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.815222 4771 scope.go:117] "RemoveContainer" containerID="739ab865b327a7cea11266222255ee0029fb2d9797cdf176b27e6f0ff2e3a925" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.817310 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.836367 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.836573 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.836743 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nn87q" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.836983 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.849866 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d8b77dfb-5n9ct"] Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.876093 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-config\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.876212 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wtgm\" (UniqueName: \"kubernetes.io/projected/7240f9bd-432f-492f-b83c-d1e265f34505-kube-api-access-4wtgm\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.876269 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.876373 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.876458 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.876545 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.877652 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.879420 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.879695 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.880397 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.880653 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-config\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.923255 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wtgm\" (UniqueName: \"kubernetes.io/projected/7240f9bd-432f-492f-b83c-d1e265f34505-kube-api-access-4wtgm\") pod \"dnsmasq-dns-5c9776ccc5-chfb4\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.978964 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-httpd-config\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.979053 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpj8d\" (UniqueName: \"kubernetes.io/projected/8f051c06-1d08-41da-9c70-8769a08d3fa1-kube-api-access-rpj8d\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.979093 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-config\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.979224 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-ovndb-tls-certs\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:40 crc kubenswrapper[4771]: I1002 10:01:40.979304 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-combined-ca-bundle\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.002201 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.082811 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-combined-ca-bundle\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.082997 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-httpd-config\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.083035 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpj8d\" (UniqueName: \"kubernetes.io/projected/8f051c06-1d08-41da-9c70-8769a08d3fa1-kube-api-access-rpj8d\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.083065 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-config\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.083104 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-ovndb-tls-certs\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.109265 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpj8d\" (UniqueName: \"kubernetes.io/projected/8f051c06-1d08-41da-9c70-8769a08d3fa1-kube-api-access-rpj8d\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.117731 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-combined-ca-bundle\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.127240 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-ovndb-tls-certs\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.127507 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-httpd-config\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.127679 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j4c5j" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="registry-server" probeResult="failure" output=< Oct 02 10:01:41 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:01:41 crc kubenswrapper[4771]: > Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.134221 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-config\") pod \"neutron-7d8b77dfb-5n9ct\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.135616 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.375823 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eb11e684-5673-479e-873b-d61930fe2d52" (UID: "eb11e684-5673-479e-873b-d61930fe2d52"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.381765 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eb11e684-5673-479e-873b-d61930fe2d52" (UID: "eb11e684-5673-479e-873b-d61930fe2d52"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.397866 4771 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.397898 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.402763 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-config" (OuterVolumeSpecName: "config") pod "eb11e684-5673-479e-873b-d61930fe2d52" (UID: "eb11e684-5673-479e-873b-d61930fe2d52"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.461348 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb11e684-5673-479e-873b-d61930fe2d52" (UID: "eb11e684-5673-479e-873b-d61930fe2d52"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.503594 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb11e684-5673-479e-873b-d61930fe2d52" (UID: "eb11e684-5673-479e-873b-d61930fe2d52"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.510481 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-nb\") pod \"eb11e684-5673-479e-873b-d61930fe2d52\" (UID: \"eb11e684-5673-479e-873b-d61930fe2d52\") " Oct 02 10:01:41 crc kubenswrapper[4771]: W1002 10:01:41.510878 4771 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/eb11e684-5673-479e-873b-d61930fe2d52/volumes/kubernetes.io~configmap/ovsdbserver-nb Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.510907 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb11e684-5673-479e-873b-d61930fe2d52" (UID: "eb11e684-5673-479e-873b-d61930fe2d52"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.518274 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.518358 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.518373 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb11e684-5673-479e-873b-d61930fe2d52-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.798444 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"55b3670c-72d0-4a3d-aa68-f0b69f9653e9","Type":"ContainerStarted","Data":"76cd1d771828373d0f8bdab9fe24252f2e5ee013bbacb9cbd6b2c2f6ce0690e5"} Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.798629 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.808691 4771 generic.go:334] "Generic (PLEG): container finished" podID="ec78b084-92a4-4eb9-ac22-bd7b51f4612c" containerID="8849634ecc2b1267f585fa7418e9281eea2feb4defe0e84bcf5a6452a9d90d36" exitCode=0 Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.808799 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svb6n" event={"ID":"ec78b084-92a4-4eb9-ac22-bd7b51f4612c","Type":"ContainerDied","Data":"8849634ecc2b1267f585fa7418e9281eea2feb4defe0e84bcf5a6452a9d90d36"} Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.827751 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.827733874 podStartE2EDuration="5.827733874s" podCreationTimestamp="2025-10-02 10:01:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:41.827172139 +0000 UTC m=+1489.474857206" watchObservedRunningTime="2025-10-02 10:01:41.827733874 +0000 UTC m=+1489.475418941" Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.859424 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-qr4df"] Oct 02 10:01:41 crc kubenswrapper[4771]: I1002 10:01:41.895211 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-qr4df"] Oct 02 10:01:42 crc kubenswrapper[4771]: I1002 10:01:42.164966 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-chfb4"] Oct 02 10:01:42 crc kubenswrapper[4771]: I1002 10:01:42.464765 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d8b77dfb-5n9ct"] Oct 02 10:01:42 crc kubenswrapper[4771]: I1002 10:01:42.895189 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d8b77dfb-5n9ct" event={"ID":"8f051c06-1d08-41da-9c70-8769a08d3fa1","Type":"ContainerStarted","Data":"a684ba0800d26c3d79eef5ee816f1fa6e219cc93dad9e8f4d9449fdcf6e392a3"} Oct 02 10:01:42 crc kubenswrapper[4771]: I1002 10:01:42.910588 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" event={"ID":"7240f9bd-432f-492f-b83c-d1e265f34505","Type":"ContainerStarted","Data":"0eacc6efbe36a57a4d20f70124e303e1f18043a2cad5db358950ab94f6374eee"} Oct 02 10:01:42 crc kubenswrapper[4771]: I1002 10:01:42.943845 4771 generic.go:334] "Generic (PLEG): container finished" podID="19df7b94-8a2c-43de-9a39-660d44eebe68" containerID="5c4bb2cb61bb8786404036fbd098013c3755be286080fc8ef6bf08caa9c69e4a" exitCode=0 Oct 02 10:01:42 crc kubenswrapper[4771]: I1002 10:01:42.944252 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"19df7b94-8a2c-43de-9a39-660d44eebe68","Type":"ContainerDied","Data":"5c4bb2cb61bb8786404036fbd098013c3755be286080fc8ef6bf08caa9c69e4a"} Oct 02 10:01:43 crc kubenswrapper[4771]: I1002 10:01:43.728193 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb11e684-5673-479e-873b-d61930fe2d52" path="/var/lib/kubelet/pods/eb11e684-5673-479e-873b-d61930fe2d52/volumes" Oct 02 10:01:43 crc kubenswrapper[4771]: I1002 10:01:43.994460 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d8b77dfb-5n9ct" event={"ID":"8f051c06-1d08-41da-9c70-8769a08d3fa1","Type":"ContainerStarted","Data":"1a83ca34e9be89357ddfe9811bc59a269810259a0885c84ee55ea230989d8975"} Oct 02 10:01:43 crc kubenswrapper[4771]: I1002 10:01:43.994507 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d8b77dfb-5n9ct" event={"ID":"8f051c06-1d08-41da-9c70-8769a08d3fa1","Type":"ContainerStarted","Data":"375d1f6c5d380d68ddb81d6180f95a3549ace994c00a2e19f4e9ab1801cc148a"} Oct 02 10:01:43 crc kubenswrapper[4771]: I1002 10:01:43.994553 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.006381 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svb6n" event={"ID":"ec78b084-92a4-4eb9-ac22-bd7b51f4612c","Type":"ContainerStarted","Data":"ebc6196e862016bec9cda955179e9950d8ed43821f551ae7940259caa344ed3f"} Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.021065 4771 generic.go:334] "Generic (PLEG): container finished" podID="7240f9bd-432f-492f-b83c-d1e265f34505" containerID="7cdee448aa91a94fdc03f59b931a6353c0e1568d3b283f1009b05d8a81f4890d" exitCode=0 Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.021498 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" event={"ID":"7240f9bd-432f-492f-b83c-d1e265f34505","Type":"ContainerDied","Data":"7cdee448aa91a94fdc03f59b931a6353c0e1568d3b283f1009b05d8a81f4890d"} Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.054036 4771 generic.go:334] "Generic (PLEG): container finished" podID="19df7b94-8a2c-43de-9a39-660d44eebe68" containerID="2a6dc00621bb08127ae6b9ba1f1abe66022c0921e7a8d72e8c2d651ec7d578ee" exitCode=0 Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.054086 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"19df7b94-8a2c-43de-9a39-660d44eebe68","Type":"ContainerDied","Data":"2a6dc00621bb08127ae6b9ba1f1abe66022c0921e7a8d72e8c2d651ec7d578ee"} Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.056875 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7d8b77dfb-5n9ct" podStartSLOduration=4.056850357 podStartE2EDuration="4.056850357s" podCreationTimestamp="2025-10-02 10:01:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:44.031906025 +0000 UTC m=+1491.679591092" watchObservedRunningTime="2025-10-02 10:01:44.056850357 +0000 UTC m=+1491.704535424" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.180990 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-66d599bfb9-7qmxw"] Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.183172 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.190786 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.191003 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.217235 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66d599bfb9-7qmxw"] Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.234583 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.373509 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f7bdb69dd-8694c" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.201:9311/healthcheck\": read tcp 10.217.0.2:33430->10.217.0.201:9311: read: connection reset by peer" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.373927 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f7bdb69dd-8694c" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.201:9311/healthcheck\": read tcp 10.217.0.2:33432->10.217.0.201:9311: read: connection reset by peer" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.391382 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-public-tls-certs\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.391590 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-httpd-config\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.392029 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvqrr\" (UniqueName: \"kubernetes.io/projected/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-kube-api-access-fvqrr\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.392667 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-ovndb-tls-certs\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.393139 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-internal-tls-certs\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.393278 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-config\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.393322 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-combined-ca-bundle\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.497480 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-internal-tls-certs\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.497546 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-config\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.497594 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-combined-ca-bundle\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.500590 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-public-tls-certs\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.501071 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-httpd-config\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.501308 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvqrr\" (UniqueName: \"kubernetes.io/projected/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-kube-api-access-fvqrr\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.501872 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-ovndb-tls-certs\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.507352 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-combined-ca-bundle\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.514927 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-public-tls-certs\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.516363 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-httpd-config\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.523229 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-ovndb-tls-certs\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.523995 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-internal-tls-certs\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.524096 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-config\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.560619 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvqrr\" (UniqueName: \"kubernetes.io/projected/fb05af7b-d804-4adf-bb0f-dedaa49a7cd0-kube-api-access-fvqrr\") pod \"neutron-66d599bfb9-7qmxw\" (UID: \"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0\") " pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.689212 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c4746c7d4-mk657" podUID="05fab4cb-323e-46a4-9b31-7e8f529e18cb" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.202:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.719391 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.809263 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-config-data\") pod \"19df7b94-8a2c-43de-9a39-660d44eebe68\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.809394 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzsg4\" (UniqueName: \"kubernetes.io/projected/19df7b94-8a2c-43de-9a39-660d44eebe68-kube-api-access-nzsg4\") pod \"19df7b94-8a2c-43de-9a39-660d44eebe68\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.809469 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-scripts\") pod \"19df7b94-8a2c-43de-9a39-660d44eebe68\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.809517 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-combined-ca-bundle\") pod \"19df7b94-8a2c-43de-9a39-660d44eebe68\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.809564 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-config-data-custom\") pod \"19df7b94-8a2c-43de-9a39-660d44eebe68\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.809675 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/19df7b94-8a2c-43de-9a39-660d44eebe68-etc-machine-id\") pod \"19df7b94-8a2c-43de-9a39-660d44eebe68\" (UID: \"19df7b94-8a2c-43de-9a39-660d44eebe68\") " Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.812353 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/19df7b94-8a2c-43de-9a39-660d44eebe68-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "19df7b94-8a2c-43de-9a39-660d44eebe68" (UID: "19df7b94-8a2c-43de-9a39-660d44eebe68"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.832344 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "19df7b94-8a2c-43de-9a39-660d44eebe68" (UID: "19df7b94-8a2c-43de-9a39-660d44eebe68"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.838051 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.845783 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19df7b94-8a2c-43de-9a39-660d44eebe68-kube-api-access-nzsg4" (OuterVolumeSpecName: "kube-api-access-nzsg4") pod "19df7b94-8a2c-43de-9a39-660d44eebe68" (UID: "19df7b94-8a2c-43de-9a39-660d44eebe68"). InnerVolumeSpecName "kube-api-access-nzsg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.857571 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-scripts" (OuterVolumeSpecName: "scripts") pod "19df7b94-8a2c-43de-9a39-660d44eebe68" (UID: "19df7b94-8a2c-43de-9a39-660d44eebe68"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.917322 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzsg4\" (UniqueName: \"kubernetes.io/projected/19df7b94-8a2c-43de-9a39-660d44eebe68-kube-api-access-nzsg4\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.917354 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.917364 4771 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:44 crc kubenswrapper[4771]: I1002 10:01:44.917373 4771 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/19df7b94-8a2c-43de-9a39-660d44eebe68-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.112286 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"19df7b94-8a2c-43de-9a39-660d44eebe68","Type":"ContainerDied","Data":"1df272aa123244b7a9fff24e9c68204f17576c71543391f6197dfa6cd9b03b0f"} Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.112554 4771 scope.go:117] "RemoveContainer" containerID="5c4bb2cb61bb8786404036fbd098013c3755be286080fc8ef6bf08caa9c69e4a" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.112333 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19df7b94-8a2c-43de-9a39-660d44eebe68" (UID: "19df7b94-8a2c-43de-9a39-660d44eebe68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.112491 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.125045 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.138888 4771 generic.go:334] "Generic (PLEG): container finished" podID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerID="f5cbb834ae1570d941b33c685c00f05ea5df87bc70a9d0e56f712cd404b3c40c" exitCode=0 Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.140627 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f7bdb69dd-8694c" event={"ID":"e17997e0-94d7-4a2f-83c1-dff9a23188cc","Type":"ContainerDied","Data":"f5cbb834ae1570d941b33c685c00f05ea5df87bc70a9d0e56f712cd404b3c40c"} Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.212806 4771 scope.go:117] "RemoveContainer" containerID="2a6dc00621bb08127ae6b9ba1f1abe66022c0921e7a8d72e8c2d651ec7d578ee" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.260381 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-config-data" (OuterVolumeSpecName: "config-data") pod "19df7b94-8a2c-43de-9a39-660d44eebe68" (UID: "19df7b94-8a2c-43de-9a39-660d44eebe68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.332732 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19df7b94-8a2c-43de-9a39-660d44eebe68-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.432359 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.498232 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.517815 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.542998 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:01:45 crc kubenswrapper[4771]: E1002 10:01:45.543534 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19df7b94-8a2c-43de-9a39-660d44eebe68" containerName="cinder-scheduler" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.543553 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="19df7b94-8a2c-43de-9a39-660d44eebe68" containerName="cinder-scheduler" Oct 02 10:01:45 crc kubenswrapper[4771]: E1002 10:01:45.543574 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerName="barbican-api-log" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.543584 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerName="barbican-api-log" Oct 02 10:01:45 crc kubenswrapper[4771]: E1002 10:01:45.543609 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerName="barbican-api" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.543616 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerName="barbican-api" Oct 02 10:01:45 crc kubenswrapper[4771]: E1002 10:01:45.543656 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19df7b94-8a2c-43de-9a39-660d44eebe68" containerName="probe" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.543662 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="19df7b94-8a2c-43de-9a39-660d44eebe68" containerName="probe" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.543860 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerName="barbican-api" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.543875 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="19df7b94-8a2c-43de-9a39-660d44eebe68" containerName="cinder-scheduler" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.543888 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" containerName="barbican-api-log" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.543899 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="19df7b94-8a2c-43de-9a39-660d44eebe68" containerName="probe" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.544550 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-combined-ca-bundle\") pod \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.544670 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-config-data\") pod \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.544757 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htw9g\" (UniqueName: \"kubernetes.io/projected/e17997e0-94d7-4a2f-83c1-dff9a23188cc-kube-api-access-htw9g\") pod \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.544811 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e17997e0-94d7-4a2f-83c1-dff9a23188cc-logs\") pod \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.544857 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-config-data-custom\") pod \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\" (UID: \"e17997e0-94d7-4a2f-83c1-dff9a23188cc\") " Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.545279 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.551374 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e17997e0-94d7-4a2f-83c1-dff9a23188cc-logs" (OuterVolumeSpecName: "logs") pod "e17997e0-94d7-4a2f-83c1-dff9a23188cc" (UID: "e17997e0-94d7-4a2f-83c1-dff9a23188cc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.551649 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e17997e0-94d7-4a2f-83c1-dff9a23188cc" (UID: "e17997e0-94d7-4a2f-83c1-dff9a23188cc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.551837 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.557468 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e17997e0-94d7-4a2f-83c1-dff9a23188cc-kube-api-access-htw9g" (OuterVolumeSpecName: "kube-api-access-htw9g") pod "e17997e0-94d7-4a2f-83c1-dff9a23188cc" (UID: "e17997e0-94d7-4a2f-83c1-dff9a23188cc"). InnerVolumeSpecName "kube-api-access-htw9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.653353 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e17997e0-94d7-4a2f-83c1-dff9a23188cc" (UID: "e17997e0-94d7-4a2f-83c1-dff9a23188cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.654505 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.654648 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.654678 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-config-data\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.654708 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.654739 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-scripts\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.654794 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h2tr\" (UniqueName: \"kubernetes.io/projected/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-kube-api-access-5h2tr\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.654982 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htw9g\" (UniqueName: \"kubernetes.io/projected/e17997e0-94d7-4a2f-83c1-dff9a23188cc-kube-api-access-htw9g\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.655005 4771 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e17997e0-94d7-4a2f-83c1-dff9a23188cc-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.655018 4771 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.655028 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.655095 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.729096 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-config-data" (OuterVolumeSpecName: "config-data") pod "e17997e0-94d7-4a2f-83c1-dff9a23188cc" (UID: "e17997e0-94d7-4a2f-83c1-dff9a23188cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.743776 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19df7b94-8a2c-43de-9a39-660d44eebe68" path="/var/lib/kubelet/pods/19df7b94-8a2c-43de-9a39-660d44eebe68/volumes" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.745391 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66d599bfb9-7qmxw"] Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.758426 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.758606 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.758641 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-config-data\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.758684 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.758717 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-scripts\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.758782 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h2tr\" (UniqueName: \"kubernetes.io/projected/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-kube-api-access-5h2tr\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.758882 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e17997e0-94d7-4a2f-83c1-dff9a23188cc-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.759858 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.765410 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-scripts\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.767452 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.772560 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-config-data\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.773326 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.793436 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h2tr\" (UniqueName: \"kubernetes.io/projected/7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61-kube-api-access-5h2tr\") pod \"cinder-scheduler-0\" (UID: \"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61\") " pod="openstack/cinder-scheduler-0" Oct 02 10:01:45 crc kubenswrapper[4771]: I1002 10:01:45.844387 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:01:46 crc kubenswrapper[4771]: I1002 10:01:46.217167 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f7bdb69dd-8694c" event={"ID":"e17997e0-94d7-4a2f-83c1-dff9a23188cc","Type":"ContainerDied","Data":"f85b46378cdbf0fa780c41165d4d193eaab1781fa5f3e2f5a351be23405e7c53"} Oct 02 10:01:46 crc kubenswrapper[4771]: I1002 10:01:46.217889 4771 scope.go:117] "RemoveContainer" containerID="f5cbb834ae1570d941b33c685c00f05ea5df87bc70a9d0e56f712cd404b3c40c" Oct 02 10:01:46 crc kubenswrapper[4771]: I1002 10:01:46.223852 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f7bdb69dd-8694c" Oct 02 10:01:46 crc kubenswrapper[4771]: I1002 10:01:46.244797 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" event={"ID":"7240f9bd-432f-492f-b83c-d1e265f34505","Type":"ContainerStarted","Data":"fe4c8fb3dd4807bf5ce5164ef52f002c755e8e09254ad70bc49cc1d953e35c5a"} Oct 02 10:01:46 crc kubenswrapper[4771]: I1002 10:01:46.281504 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f7bdb69dd-8694c"] Oct 02 10:01:46 crc kubenswrapper[4771]: I1002 10:01:46.282105 4771 scope.go:117] "RemoveContainer" containerID="83baa0af28908c2e64eec13434197211690a854fe2927d4cc4a5901c326094f2" Oct 02 10:01:46 crc kubenswrapper[4771]: I1002 10:01:46.299060 4771 generic.go:334] "Generic (PLEG): container finished" podID="ec78b084-92a4-4eb9-ac22-bd7b51f4612c" containerID="ebc6196e862016bec9cda955179e9950d8ed43821f551ae7940259caa344ed3f" exitCode=0 Oct 02 10:01:46 crc kubenswrapper[4771]: I1002 10:01:46.299333 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svb6n" event={"ID":"ec78b084-92a4-4eb9-ac22-bd7b51f4612c","Type":"ContainerDied","Data":"ebc6196e862016bec9cda955179e9950d8ed43821f551ae7940259caa344ed3f"} Oct 02 10:01:46 crc kubenswrapper[4771]: I1002 10:01:46.304292 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66d599bfb9-7qmxw" event={"ID":"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0","Type":"ContainerStarted","Data":"e6fc34fdf6316eb79765e420116b8239b7aa89d8734020776ac086d859699752"} Oct 02 10:01:46 crc kubenswrapper[4771]: I1002 10:01:46.304358 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66d599bfb9-7qmxw" event={"ID":"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0","Type":"ContainerStarted","Data":"204eb4390fdfcdafb673a083d8c13513041f72332d5b9cda442cde252c22e707"} Oct 02 10:01:46 crc kubenswrapper[4771]: I1002 10:01:46.315919 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5f7bdb69dd-8694c"] Oct 02 10:01:46 crc kubenswrapper[4771]: I1002 10:01:46.327830 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" podStartSLOduration=6.327802529 podStartE2EDuration="6.327802529s" podCreationTimestamp="2025-10-02 10:01:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:46.282657342 +0000 UTC m=+1493.930342429" watchObservedRunningTime="2025-10-02 10:01:46.327802529 +0000 UTC m=+1493.975487596" Oct 02 10:01:46 crc kubenswrapper[4771]: W1002 10:01:46.710951 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e1d5ebb_a7e6_4509_bd86_f2fa34df2b61.slice/crio-c1f0966563c89de2ccc2a3d79b26c188dcf35edda0277c4ddfa9ce7d58d7a672 WatchSource:0}: Error finding container c1f0966563c89de2ccc2a3d79b26c188dcf35edda0277c4ddfa9ce7d58d7a672: Status 404 returned error can't find the container with id c1f0966563c89de2ccc2a3d79b26c188dcf35edda0277c4ddfa9ce7d58d7a672 Oct 02 10:01:46 crc kubenswrapper[4771]: I1002 10:01:46.732526 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:01:47 crc kubenswrapper[4771]: I1002 10:01:47.355318 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66d599bfb9-7qmxw" event={"ID":"fb05af7b-d804-4adf-bb0f-dedaa49a7cd0","Type":"ContainerStarted","Data":"e0b4f03e26297595c08c3a3e505b5480ef2ac8c1dda8194029eaf18a367e6fef"} Oct 02 10:01:47 crc kubenswrapper[4771]: I1002 10:01:47.362508 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:01:47 crc kubenswrapper[4771]: I1002 10:01:47.385556 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-66d599bfb9-7qmxw" podStartSLOduration=3.385532948 podStartE2EDuration="3.385532948s" podCreationTimestamp="2025-10-02 10:01:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:47.379162996 +0000 UTC m=+1495.026848063" watchObservedRunningTime="2025-10-02 10:01:47.385532948 +0000 UTC m=+1495.033218025" Oct 02 10:01:47 crc kubenswrapper[4771]: I1002 10:01:47.390931 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61","Type":"ContainerStarted","Data":"c1f0966563c89de2ccc2a3d79b26c188dcf35edda0277c4ddfa9ce7d58d7a672"} Oct 02 10:01:47 crc kubenswrapper[4771]: I1002 10:01:47.391026 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:47 crc kubenswrapper[4771]: I1002 10:01:47.722526 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e17997e0-94d7-4a2f-83c1-dff9a23188cc" path="/var/lib/kubelet/pods/e17997e0-94d7-4a2f-83c1-dff9a23188cc/volumes" Oct 02 10:01:48 crc kubenswrapper[4771]: I1002 10:01:48.413269 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61","Type":"ContainerStarted","Data":"fb16b672f98e7e7de6319e4d75aa65872e8f7c0ff381500a4e2c60959e11988b"} Oct 02 10:01:48 crc kubenswrapper[4771]: I1002 10:01:48.435922 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svb6n" event={"ID":"ec78b084-92a4-4eb9-ac22-bd7b51f4612c","Type":"ContainerStarted","Data":"4dd862c9c342a8db8e0fcaf6b368ca9bb165040c0c6a5f3c28edd84b62c30bdc"} Oct 02 10:01:48 crc kubenswrapper[4771]: I1002 10:01:48.468914 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-svb6n" podStartSLOduration=6.007872173 podStartE2EDuration="11.468888259s" podCreationTimestamp="2025-10-02 10:01:37 +0000 UTC" firstStartedPulling="2025-10-02 10:01:41.835987926 +0000 UTC m=+1489.483672993" lastFinishedPulling="2025-10-02 10:01:47.297004012 +0000 UTC m=+1494.944689079" observedRunningTime="2025-10-02 10:01:48.456268599 +0000 UTC m=+1496.103953686" watchObservedRunningTime="2025-10-02 10:01:48.468888259 +0000 UTC m=+1496.116573326" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.063733 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.064168 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-58587f9544-7mwm6" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.493352 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61","Type":"ContainerStarted","Data":"2cf23c85cfd248cac673663c3f6e375e417f400e1b27b27def0f26e7610f3c9a"} Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.536525 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.536496795 podStartE2EDuration="4.536496795s" podCreationTimestamp="2025-10-02 10:01:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:49.509779985 +0000 UTC m=+1497.157465062" watchObservedRunningTime="2025-10-02 10:01:49.536496795 +0000 UTC m=+1497.184181862" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.640699 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-668c9ddd79-9f298"] Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.643544 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.650206 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.650474 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.660936 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.664991 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-668c9ddd79-9f298"] Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.699324 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/285035a2-e600-4571-9462-e029a84b3779-internal-tls-certs\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.699872 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285035a2-e600-4571-9462-e029a84b3779-config-data\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.699988 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285035a2-e600-4571-9462-e029a84b3779-combined-ca-bundle\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.700032 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv6s2\" (UniqueName: \"kubernetes.io/projected/285035a2-e600-4571-9462-e029a84b3779-kube-api-access-rv6s2\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.700324 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/285035a2-e600-4571-9462-e029a84b3779-run-httpd\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.700352 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/285035a2-e600-4571-9462-e029a84b3779-public-tls-certs\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.700470 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/285035a2-e600-4571-9462-e029a84b3779-etc-swift\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.700509 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/285035a2-e600-4571-9462-e029a84b3779-log-httpd\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.802852 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285035a2-e600-4571-9462-e029a84b3779-config-data\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.803003 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285035a2-e600-4571-9462-e029a84b3779-combined-ca-bundle\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.803046 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv6s2\" (UniqueName: \"kubernetes.io/projected/285035a2-e600-4571-9462-e029a84b3779-kube-api-access-rv6s2\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.803430 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/285035a2-e600-4571-9462-e029a84b3779-run-httpd\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.803466 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/285035a2-e600-4571-9462-e029a84b3779-public-tls-certs\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.803535 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/285035a2-e600-4571-9462-e029a84b3779-etc-swift\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.803559 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/285035a2-e600-4571-9462-e029a84b3779-log-httpd\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.803627 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/285035a2-e600-4571-9462-e029a84b3779-internal-tls-certs\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.805715 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/285035a2-e600-4571-9462-e029a84b3779-log-httpd\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.806110 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/285035a2-e600-4571-9462-e029a84b3779-run-httpd\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.811541 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/285035a2-e600-4571-9462-e029a84b3779-public-tls-certs\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.812926 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/285035a2-e600-4571-9462-e029a84b3779-internal-tls-certs\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.814276 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/285035a2-e600-4571-9462-e029a84b3779-etc-swift\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.814730 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285035a2-e600-4571-9462-e029a84b3779-config-data\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.826116 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv6s2\" (UniqueName: \"kubernetes.io/projected/285035a2-e600-4571-9462-e029a84b3779-kube-api-access-rv6s2\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.851375 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285035a2-e600-4571-9462-e029a84b3779-combined-ca-bundle\") pod \"swift-proxy-668c9ddd79-9f298\" (UID: \"285035a2-e600-4571-9462-e029a84b3779\") " pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.967036 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-649769f8cf-k8cph"] Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.968727 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.974017 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.980707 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-4fps8" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.982215 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.985723 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Oct 02 10:01:49 crc kubenswrapper[4771]: I1002 10:01:49.997375 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-649769f8cf-k8cph"] Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.029808 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-combined-ca-bundle\") pod \"heat-engine-649769f8cf-k8cph\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.038834 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-config-data-custom\") pod \"heat-engine-649769f8cf-k8cph\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.039119 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf9z8\" (UniqueName: \"kubernetes.io/projected/d7b40340-a806-490f-9842-a074ffba2e93-kube-api-access-tf9z8\") pod \"heat-engine-649769f8cf-k8cph\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.039268 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-config-data\") pod \"heat-engine-649769f8cf-k8cph\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.153567 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf9z8\" (UniqueName: \"kubernetes.io/projected/d7b40340-a806-490f-9842-a074ffba2e93-kube-api-access-tf9z8\") pod \"heat-engine-649769f8cf-k8cph\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.153701 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-config-data\") pod \"heat-engine-649769f8cf-k8cph\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.153871 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-combined-ca-bundle\") pod \"heat-engine-649769f8cf-k8cph\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.154144 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-config-data-custom\") pod \"heat-engine-649769f8cf-k8cph\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.169727 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-chfb4"] Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.170011 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" podUID="7240f9bd-432f-492f-b83c-d1e265f34505" containerName="dnsmasq-dns" containerID="cri-o://fe4c8fb3dd4807bf5ce5164ef52f002c755e8e09254ad70bc49cc1d953e35c5a" gracePeriod=10 Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.172851 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.173172 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-combined-ca-bundle\") pod \"heat-engine-649769f8cf-k8cph\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.181942 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-config-data-custom\") pod \"heat-engine-649769f8cf-k8cph\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.189726 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.195539 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-config-data\") pod \"heat-engine-649769f8cf-k8cph\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.214111 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf9z8\" (UniqueName: \"kubernetes.io/projected/d7b40340-a806-490f-9842-a074ffba2e93-kube-api-access-tf9z8\") pod \"heat-engine-649769f8cf-k8cph\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.307176 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-8xp24"] Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.313194 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.329459 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-8xp24"] Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.332323 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.418752 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-c7b979484-9sfpc"] Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.424632 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.433826 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.437469 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.457201 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-c7b979484-9sfpc"] Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.466988 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.467057 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l94dk\" (UniqueName: \"kubernetes.io/projected/0ec19f19-2184-43e0-b683-d35db4a08da0-kube-api-access-l94dk\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.467087 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.467145 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.467193 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.467417 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-config\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.496842 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-88f48768-bxk7h"] Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.498641 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.501945 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.525075 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-88f48768-bxk7h"] Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.583535 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-combined-ca-bundle\") pod \"heat-api-88f48768-bxk7h\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.587252 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-config\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.587410 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsvc2\" (UniqueName: \"kubernetes.io/projected/49ac894e-c59d-40e0-94e0-37c0913afd0f-kube-api-access-wsvc2\") pod \"heat-api-88f48768-bxk7h\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.587639 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-config-data-custom\") pod \"heat-cfnapi-c7b979484-9sfpc\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.587878 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-config-data-custom\") pod \"heat-api-88f48768-bxk7h\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.587903 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmq68\" (UniqueName: \"kubernetes.io/projected/9404916e-870c-403e-9320-25fda656bf6a-kube-api-access-wmq68\") pod \"heat-cfnapi-c7b979484-9sfpc\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.587996 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-config-data\") pod \"heat-cfnapi-c7b979484-9sfpc\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.588183 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.588322 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l94dk\" (UniqueName: \"kubernetes.io/projected/0ec19f19-2184-43e0-b683-d35db4a08da0-kube-api-access-l94dk\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.588350 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.588426 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.588457 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-config-data\") pod \"heat-api-88f48768-bxk7h\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.588699 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.588739 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-combined-ca-bundle\") pod \"heat-cfnapi-c7b979484-9sfpc\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.590889 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.590913 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.591572 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.592108 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.597699 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-config\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.616974 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l94dk\" (UniqueName: \"kubernetes.io/projected/0ec19f19-2184-43e0-b683-d35db4a08da0-kube-api-access-l94dk\") pod \"dnsmasq-dns-7756b9d78c-8xp24\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.630262 4771 generic.go:334] "Generic (PLEG): container finished" podID="7240f9bd-432f-492f-b83c-d1e265f34505" containerID="fe4c8fb3dd4807bf5ce5164ef52f002c755e8e09254ad70bc49cc1d953e35c5a" exitCode=0 Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.631632 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" event={"ID":"7240f9bd-432f-492f-b83c-d1e265f34505","Type":"ContainerDied","Data":"fe4c8fb3dd4807bf5ce5164ef52f002c755e8e09254ad70bc49cc1d953e35c5a"} Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.676045 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j4c5j"] Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.694267 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsvc2\" (UniqueName: \"kubernetes.io/projected/49ac894e-c59d-40e0-94e0-37c0913afd0f-kube-api-access-wsvc2\") pod \"heat-api-88f48768-bxk7h\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.694568 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-config-data-custom\") pod \"heat-cfnapi-c7b979484-9sfpc\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.694720 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-config-data-custom\") pod \"heat-api-88f48768-bxk7h\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.694793 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmq68\" (UniqueName: \"kubernetes.io/projected/9404916e-870c-403e-9320-25fda656bf6a-kube-api-access-wmq68\") pod \"heat-cfnapi-c7b979484-9sfpc\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.694873 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-config-data\") pod \"heat-cfnapi-c7b979484-9sfpc\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.695005 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-config-data\") pod \"heat-api-88f48768-bxk7h\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.695106 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-combined-ca-bundle\") pod \"heat-cfnapi-c7b979484-9sfpc\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.695226 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-combined-ca-bundle\") pod \"heat-api-88f48768-bxk7h\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.707440 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-config-data\") pod \"heat-cfnapi-c7b979484-9sfpc\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.707816 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.712436 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-config-data-custom\") pod \"heat-api-88f48768-bxk7h\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.716296 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-config-data-custom\") pod \"heat-cfnapi-c7b979484-9sfpc\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.724901 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-combined-ca-bundle\") pod \"heat-cfnapi-c7b979484-9sfpc\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.727724 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-config-data\") pod \"heat-api-88f48768-bxk7h\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.745666 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmq68\" (UniqueName: \"kubernetes.io/projected/9404916e-870c-403e-9320-25fda656bf6a-kube-api-access-wmq68\") pod \"heat-cfnapi-c7b979484-9sfpc\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.746885 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-combined-ca-bundle\") pod \"heat-api-88f48768-bxk7h\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.766976 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsvc2\" (UniqueName: \"kubernetes.io/projected/49ac894e-c59d-40e0-94e0-37c0913afd0f-kube-api-access-wsvc2\") pod \"heat-api-88f48768-bxk7h\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.788286 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.844518 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:01:50 crc kubenswrapper[4771]: I1002 10:01:50.846534 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 10:01:51 crc kubenswrapper[4771]: I1002 10:01:51.004519 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" podUID="7240f9bd-432f-492f-b83c-d1e265f34505" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.209:5353: connect: connection refused" Oct 02 10:01:51 crc kubenswrapper[4771]: I1002 10:01:51.556746 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-668c9ddd79-9f298"] Oct 02 10:01:51 crc kubenswrapper[4771]: I1002 10:01:51.701760 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j4c5j" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="registry-server" containerID="cri-o://9dd20d380201ee099ec8bcaf1a7d049eb27504b670bd8087bb70301b7a4509b0" gracePeriod=2 Oct 02 10:01:51 crc kubenswrapper[4771]: I1002 10:01:51.727664 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-668c9ddd79-9f298" event={"ID":"285035a2-e600-4571-9462-e029a84b3779","Type":"ContainerStarted","Data":"7ff84a05a9161169400aaf7248cbeb57f96f0308310bbdb00944d882169a2feb"} Oct 02 10:01:51 crc kubenswrapper[4771]: I1002 10:01:51.929892 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="55b3670c-72d0-4a3d-aa68-f0b69f9653e9" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.206:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.203681 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.310635 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-dns-swift-storage-0\") pod \"7240f9bd-432f-492f-b83c-d1e265f34505\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.310722 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-ovsdbserver-sb\") pod \"7240f9bd-432f-492f-b83c-d1e265f34505\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.310749 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wtgm\" (UniqueName: \"kubernetes.io/projected/7240f9bd-432f-492f-b83c-d1e265f34505-kube-api-access-4wtgm\") pod \"7240f9bd-432f-492f-b83c-d1e265f34505\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.310793 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-ovsdbserver-nb\") pod \"7240f9bd-432f-492f-b83c-d1e265f34505\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.310958 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-config\") pod \"7240f9bd-432f-492f-b83c-d1e265f34505\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.327340 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-dns-svc\") pod \"7240f9bd-432f-492f-b83c-d1e265f34505\" (UID: \"7240f9bd-432f-492f-b83c-d1e265f34505\") " Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.359435 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7240f9bd-432f-492f-b83c-d1e265f34505-kube-api-access-4wtgm" (OuterVolumeSpecName: "kube-api-access-4wtgm") pod "7240f9bd-432f-492f-b83c-d1e265f34505" (UID: "7240f9bd-432f-492f-b83c-d1e265f34505"). InnerVolumeSpecName "kube-api-access-4wtgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.436155 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wtgm\" (UniqueName: \"kubernetes.io/projected/7240f9bd-432f-492f-b83c-d1e265f34505-kube-api-access-4wtgm\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:52 crc kubenswrapper[4771]: W1002 10:01:52.500841 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7b40340_a806_490f_9842_a074ffba2e93.slice/crio-04990be5940d31e6b9fe743a99f836720a02dcd7bd81ce0a745c3a5c448c5ef9 WatchSource:0}: Error finding container 04990be5940d31e6b9fe743a99f836720a02dcd7bd81ce0a745c3a5c448c5ef9: Status 404 returned error can't find the container with id 04990be5940d31e6b9fe743a99f836720a02dcd7bd81ce0a745c3a5c448c5ef9 Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.509099 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-649769f8cf-k8cph"] Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.552418 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-8xp24"] Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.582262 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-config" (OuterVolumeSpecName: "config") pod "7240f9bd-432f-492f-b83c-d1e265f34505" (UID: "7240f9bd-432f-492f-b83c-d1e265f34505"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.593389 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-c7b979484-9sfpc"] Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.611828 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7240f9bd-432f-492f-b83c-d1e265f34505" (UID: "7240f9bd-432f-492f-b83c-d1e265f34505"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.651567 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.651602 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.663761 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7240f9bd-432f-492f-b83c-d1e265f34505" (UID: "7240f9bd-432f-492f-b83c-d1e265f34505"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.663775 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7240f9bd-432f-492f-b83c-d1e265f34505" (UID: "7240f9bd-432f-492f-b83c-d1e265f34505"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.664433 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7240f9bd-432f-492f-b83c-d1e265f34505" (UID: "7240f9bd-432f-492f-b83c-d1e265f34505"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.739916 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" event={"ID":"0ec19f19-2184-43e0-b683-d35db4a08da0","Type":"ContainerStarted","Data":"f987c0fe29646178d6c930a015339d9e525eba607468582c0b206d53937fe2fe"} Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.759425 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.759454 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.759466 4771 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7240f9bd-432f-492f-b83c-d1e265f34505-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.788391 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-c7b979484-9sfpc" event={"ID":"9404916e-870c-403e-9320-25fda656bf6a","Type":"ContainerStarted","Data":"28916e258dc641e41144e4e82834d76077acaa926d2e92258a6a4e294d834abf"} Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.809446 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-88f48768-bxk7h"] Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.876547 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" event={"ID":"7240f9bd-432f-492f-b83c-d1e265f34505","Type":"ContainerDied","Data":"0eacc6efbe36a57a4d20f70124e303e1f18043a2cad5db358950ab94f6374eee"} Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.876618 4771 scope.go:117] "RemoveContainer" containerID="fe4c8fb3dd4807bf5ce5164ef52f002c755e8e09254ad70bc49cc1d953e35c5a" Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.876817 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-chfb4" Oct 02 10:01:52 crc kubenswrapper[4771]: W1002 10:01:52.931099 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49ac894e_c59d_40e0_94e0_37c0913afd0f.slice/crio-556b37d3633a6fa229aaaf1ef92b3ca46075ae7bbd04e687b790fa8b0beaeb7d WatchSource:0}: Error finding container 556b37d3633a6fa229aaaf1ef92b3ca46075ae7bbd04e687b790fa8b0beaeb7d: Status 404 returned error can't find the container with id 556b37d3633a6fa229aaaf1ef92b3ca46075ae7bbd04e687b790fa8b0beaeb7d Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.931166 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-668c9ddd79-9f298" event={"ID":"285035a2-e600-4571-9462-e029a84b3779","Type":"ContainerStarted","Data":"8ab77428694547086f6e0b1449d981e5a184253eb457973739db2fd5d7e72581"} Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.960753 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-649769f8cf-k8cph" event={"ID":"d7b40340-a806-490f-9842-a074ffba2e93","Type":"ContainerStarted","Data":"04990be5940d31e6b9fe743a99f836720a02dcd7bd81ce0a745c3a5c448c5ef9"} Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.987616 4771 generic.go:334] "Generic (PLEG): container finished" podID="3184235b-9379-4b54-803d-dbb52582e06b" containerID="9dd20d380201ee099ec8bcaf1a7d049eb27504b670bd8087bb70301b7a4509b0" exitCode=0 Oct 02 10:01:52 crc kubenswrapper[4771]: I1002 10:01:52.987683 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4c5j" event={"ID":"3184235b-9379-4b54-803d-dbb52582e06b","Type":"ContainerDied","Data":"9dd20d380201ee099ec8bcaf1a7d049eb27504b670bd8087bb70301b7a4509b0"} Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.017945 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-chfb4"] Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.047781 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-chfb4"] Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.195621 4771 scope.go:117] "RemoveContainer" containerID="7cdee448aa91a94fdc03f59b931a6353c0e1568d3b283f1009b05d8a81f4890d" Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.392602 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.407908 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3184235b-9379-4b54-803d-dbb52582e06b-utilities\") pod \"3184235b-9379-4b54-803d-dbb52582e06b\" (UID: \"3184235b-9379-4b54-803d-dbb52582e06b\") " Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.407963 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfm2f\" (UniqueName: \"kubernetes.io/projected/3184235b-9379-4b54-803d-dbb52582e06b-kube-api-access-tfm2f\") pod \"3184235b-9379-4b54-803d-dbb52582e06b\" (UID: \"3184235b-9379-4b54-803d-dbb52582e06b\") " Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.407986 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3184235b-9379-4b54-803d-dbb52582e06b-catalog-content\") pod \"3184235b-9379-4b54-803d-dbb52582e06b\" (UID: \"3184235b-9379-4b54-803d-dbb52582e06b\") " Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.408698 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3184235b-9379-4b54-803d-dbb52582e06b-utilities" (OuterVolumeSpecName: "utilities") pod "3184235b-9379-4b54-803d-dbb52582e06b" (UID: "3184235b-9379-4b54-803d-dbb52582e06b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.434194 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3184235b-9379-4b54-803d-dbb52582e06b-kube-api-access-tfm2f" (OuterVolumeSpecName: "kube-api-access-tfm2f") pod "3184235b-9379-4b54-803d-dbb52582e06b" (UID: "3184235b-9379-4b54-803d-dbb52582e06b"). InnerVolumeSpecName "kube-api-access-tfm2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.507741 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.513343 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="ceilometer-central-agent" containerID="cri-o://74bf1bb5db2b07de6f282f2c7a2c8ac004d12d5baae3d4580a58192f6649d863" gracePeriod=30 Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.511084 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3184235b-9379-4b54-803d-dbb52582e06b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.513832 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfm2f\" (UniqueName: \"kubernetes.io/projected/3184235b-9379-4b54-803d-dbb52582e06b-kube-api-access-tfm2f\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.515549 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="proxy-httpd" containerID="cri-o://bb7ea1b59d47aa056d3f4df5e6ca203ac5091bfe041d713a62b00549177eb312" gracePeriod=30 Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.515754 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="sg-core" containerID="cri-o://48ffa7722cc5a44891b25a0bf45bdb35e6a21d55b848def01947b2101bd08f22" gracePeriod=30 Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.515881 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="ceilometer-notification-agent" containerID="cri-o://00e4574206a45d648bcda5196849131bf2c3fb95fbf0397513229cc9a789de07" gracePeriod=30 Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.582114 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3184235b-9379-4b54-803d-dbb52582e06b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3184235b-9379-4b54-803d-dbb52582e06b" (UID: "3184235b-9379-4b54-803d-dbb52582e06b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.632094 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3184235b-9379-4b54-803d-dbb52582e06b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.739211 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7240f9bd-432f-492f-b83c-d1e265f34505" path="/var/lib/kubelet/pods/7240f9bd-432f-492f-b83c-d1e265f34505/volumes" Oct 02 10:01:53 crc kubenswrapper[4771]: I1002 10:01:53.960299 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="55b3670c-72d0-4a3d-aa68-f0b69f9653e9" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.206:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.137567 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-668c9ddd79-9f298" event={"ID":"285035a2-e600-4571-9462-e029a84b3779","Type":"ContainerStarted","Data":"cd36a8f78d2db8d9134b46fee182da6a8b9138699e6b38b15054c36c4021d960"} Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.139302 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.139343 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.175036 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-649769f8cf-k8cph" event={"ID":"d7b40340-a806-490f-9842-a074ffba2e93","Type":"ContainerStarted","Data":"d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98"} Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.175310 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.192105 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-668c9ddd79-9f298" podStartSLOduration=5.192081751 podStartE2EDuration="5.192081751s" podCreationTimestamp="2025-10-02 10:01:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:54.166767899 +0000 UTC m=+1501.814452976" watchObservedRunningTime="2025-10-02 10:01:54.192081751 +0000 UTC m=+1501.839766818" Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.219701 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4c5j" event={"ID":"3184235b-9379-4b54-803d-dbb52582e06b","Type":"ContainerDied","Data":"8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f"} Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.219763 4771 scope.go:117] "RemoveContainer" containerID="9dd20d380201ee099ec8bcaf1a7d049eb27504b670bd8087bb70301b7a4509b0" Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.219961 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j4c5j" Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.247274 4771 generic.go:334] "Generic (PLEG): container finished" podID="0ec19f19-2184-43e0-b683-d35db4a08da0" containerID="ef2ceb84f86437104a92e490ba4ddae48fcb8417a02b16deac0e333d9b6aaaa7" exitCode=0 Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.247358 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" event={"ID":"0ec19f19-2184-43e0-b683-d35db4a08da0","Type":"ContainerDied","Data":"ef2ceb84f86437104a92e490ba4ddae48fcb8417a02b16deac0e333d9b6aaaa7"} Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.253005 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-649769f8cf-k8cph" podStartSLOduration=5.252981902 podStartE2EDuration="5.252981902s" podCreationTimestamp="2025-10-02 10:01:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:54.228929594 +0000 UTC m=+1501.876614671" watchObservedRunningTime="2025-10-02 10:01:54.252981902 +0000 UTC m=+1501.900666959" Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.270262 4771 generic.go:334] "Generic (PLEG): container finished" podID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerID="bb7ea1b59d47aa056d3f4df5e6ca203ac5091bfe041d713a62b00549177eb312" exitCode=0 Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.270304 4771 generic.go:334] "Generic (PLEG): container finished" podID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerID="48ffa7722cc5a44891b25a0bf45bdb35e6a21d55b848def01947b2101bd08f22" exitCode=2 Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.270360 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f","Type":"ContainerDied","Data":"bb7ea1b59d47aa056d3f4df5e6ca203ac5091bfe041d713a62b00549177eb312"} Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.270397 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f","Type":"ContainerDied","Data":"48ffa7722cc5a44891b25a0bf45bdb35e6a21d55b848def01947b2101bd08f22"} Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.295180 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-88f48768-bxk7h" event={"ID":"49ac894e-c59d-40e0-94e0-37c0913afd0f","Type":"ContainerStarted","Data":"556b37d3633a6fa229aaaf1ef92b3ca46075ae7bbd04e687b790fa8b0beaeb7d"} Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.358206 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j4c5j"] Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.383947 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j4c5j"] Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.525112 4771 scope.go:117] "RemoveContainer" containerID="c4feb8e59691a866aa029f4ac85d391f25603c66aeefe0bae491495d70f46b87" Oct 02 10:01:54 crc kubenswrapper[4771]: I1002 10:01:54.637635 4771 scope.go:117] "RemoveContainer" containerID="4b23ae1a19066feda4479e04b20e05bd58683722fdfcfb9179f7f69c739fb215" Oct 02 10:01:55 crc kubenswrapper[4771]: I1002 10:01:55.392689 4771 generic.go:334] "Generic (PLEG): container finished" podID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerID="74bf1bb5db2b07de6f282f2c7a2c8ac004d12d5baae3d4580a58192f6649d863" exitCode=0 Oct 02 10:01:55 crc kubenswrapper[4771]: I1002 10:01:55.392775 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f","Type":"ContainerDied","Data":"74bf1bb5db2b07de6f282f2c7a2c8ac004d12d5baae3d4580a58192f6649d863"} Oct 02 10:01:55 crc kubenswrapper[4771]: I1002 10:01:55.435001 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" event={"ID":"0ec19f19-2184-43e0-b683-d35db4a08da0","Type":"ContainerStarted","Data":"4c75700c18b034995223129b8b2d3e4c4146fae13fe3a9e7558ab848cc004eff"} Oct 02 10:01:55 crc kubenswrapper[4771]: I1002 10:01:55.435240 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:01:55 crc kubenswrapper[4771]: I1002 10:01:55.474146 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" podStartSLOduration=5.474100577 podStartE2EDuration="5.474100577s" podCreationTimestamp="2025-10-02 10:01:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:55.469871483 +0000 UTC m=+1503.117556550" watchObservedRunningTime="2025-10-02 10:01:55.474100577 +0000 UTC m=+1503.121785644" Oct 02 10:01:55 crc kubenswrapper[4771]: I1002 10:01:55.720073 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3184235b-9379-4b54-803d-dbb52582e06b" path="/var/lib/kubelet/pods/3184235b-9379-4b54-803d-dbb52582e06b/volumes" Oct 02 10:01:56 crc kubenswrapper[4771]: I1002 10:01:56.944358 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="55b3670c-72d0-4a3d-aa68-f0b69f9653e9" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.206:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 10:01:57 crc kubenswrapper[4771]: I1002 10:01:57.053294 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 10:01:57 crc kubenswrapper[4771]: E1002 10:01:57.078155 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice/crio-8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f\": RecentStats: unable to find data in memory cache]" Oct 02 10:01:57 crc kubenswrapper[4771]: I1002 10:01:57.277676 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 10:01:58 crc kubenswrapper[4771]: I1002 10:01:58.130146 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:58 crc kubenswrapper[4771]: I1002 10:01:58.130691 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:58 crc kubenswrapper[4771]: I1002 10:01:58.264540 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:58 crc kubenswrapper[4771]: I1002 10:01:58.606484 4771 generic.go:334] "Generic (PLEG): container finished" podID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerID="00e4574206a45d648bcda5196849131bf2c3fb95fbf0397513229cc9a789de07" exitCode=0 Oct 02 10:01:58 crc kubenswrapper[4771]: I1002 10:01:58.606840 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f","Type":"ContainerDied","Data":"00e4574206a45d648bcda5196849131bf2c3fb95fbf0397513229cc9a789de07"} Oct 02 10:01:58 crc kubenswrapper[4771]: I1002 10:01:58.683855 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:01:58 crc kubenswrapper[4771]: I1002 10:01:58.760627 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svb6n"] Oct 02 10:01:58 crc kubenswrapper[4771]: I1002 10:01:58.977168 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="55b3670c-72d0-4a3d-aa68-f0b69f9653e9" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.206:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.187817 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-868b7bfb89-mjrmr"] Oct 02 10:01:59 crc kubenswrapper[4771]: E1002 10:01:59.188424 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="extract-content" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.188443 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="extract-content" Oct 02 10:01:59 crc kubenswrapper[4771]: E1002 10:01:59.188468 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7240f9bd-432f-492f-b83c-d1e265f34505" containerName="dnsmasq-dns" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.188475 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="7240f9bd-432f-492f-b83c-d1e265f34505" containerName="dnsmasq-dns" Oct 02 10:01:59 crc kubenswrapper[4771]: E1002 10:01:59.188509 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="extract-utilities" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.188516 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="extract-utilities" Oct 02 10:01:59 crc kubenswrapper[4771]: E1002 10:01:59.188526 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="registry-server" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.188532 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="registry-server" Oct 02 10:01:59 crc kubenswrapper[4771]: E1002 10:01:59.188549 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7240f9bd-432f-492f-b83c-d1e265f34505" containerName="init" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.188555 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="7240f9bd-432f-492f-b83c-d1e265f34505" containerName="init" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.188782 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="3184235b-9379-4b54-803d-dbb52582e06b" containerName="registry-server" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.188814 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="7240f9bd-432f-492f-b83c-d1e265f34505" containerName="dnsmasq-dns" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.189777 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.212336 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-868b7bfb89-mjrmr"] Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.230934 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drwsr\" (UniqueName: \"kubernetes.io/projected/a4bc89da-2332-412a-b7af-6056f830310a-kube-api-access-drwsr\") pod \"heat-engine-868b7bfb89-mjrmr\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.231074 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-config-data-custom\") pod \"heat-engine-868b7bfb89-mjrmr\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.231114 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-combined-ca-bundle\") pod \"heat-engine-868b7bfb89-mjrmr\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.231202 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-config-data\") pod \"heat-engine-868b7bfb89-mjrmr\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.234258 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-8ff6c7648-wtgnm"] Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.237589 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.246919 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5d49fd5cb-kh9dz"] Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.248966 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.260229 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-8ff6c7648-wtgnm"] Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.281607 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5d49fd5cb-kh9dz"] Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.333977 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drwsr\" (UniqueName: \"kubernetes.io/projected/a4bc89da-2332-412a-b7af-6056f830310a-kube-api-access-drwsr\") pod \"heat-engine-868b7bfb89-mjrmr\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.334447 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlzgp\" (UniqueName: \"kubernetes.io/projected/ad488e84-ee1a-47f8-b626-fdc87919520d-kube-api-access-vlzgp\") pod \"heat-cfnapi-5d49fd5cb-kh9dz\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.334648 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-config-data-custom\") pod \"heat-cfnapi-5d49fd5cb-kh9dz\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.334928 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s47k6\" (UniqueName: \"kubernetes.io/projected/983d8aca-7ee4-4105-9678-63d608645975-kube-api-access-s47k6\") pod \"heat-api-8ff6c7648-wtgnm\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.335099 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-combined-ca-bundle\") pod \"heat-cfnapi-5d49fd5cb-kh9dz\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.335256 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-config-data-custom\") pod \"heat-api-8ff6c7648-wtgnm\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.335400 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-config-data-custom\") pod \"heat-engine-868b7bfb89-mjrmr\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.338572 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-combined-ca-bundle\") pod \"heat-engine-868b7bfb89-mjrmr\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.338718 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-config-data\") pod \"heat-api-8ff6c7648-wtgnm\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.338969 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-config-data\") pod \"heat-engine-868b7bfb89-mjrmr\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.339159 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-combined-ca-bundle\") pod \"heat-api-8ff6c7648-wtgnm\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.339305 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-config-data\") pod \"heat-cfnapi-5d49fd5cb-kh9dz\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.350013 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-config-data-custom\") pod \"heat-engine-868b7bfb89-mjrmr\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.352561 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-config-data\") pod \"heat-engine-868b7bfb89-mjrmr\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.364000 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drwsr\" (UniqueName: \"kubernetes.io/projected/a4bc89da-2332-412a-b7af-6056f830310a-kube-api-access-drwsr\") pod \"heat-engine-868b7bfb89-mjrmr\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.375095 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-combined-ca-bundle\") pod \"heat-engine-868b7bfb89-mjrmr\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.443517 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlzgp\" (UniqueName: \"kubernetes.io/projected/ad488e84-ee1a-47f8-b626-fdc87919520d-kube-api-access-vlzgp\") pod \"heat-cfnapi-5d49fd5cb-kh9dz\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.444184 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-config-data-custom\") pod \"heat-cfnapi-5d49fd5cb-kh9dz\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.445023 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s47k6\" (UniqueName: \"kubernetes.io/projected/983d8aca-7ee4-4105-9678-63d608645975-kube-api-access-s47k6\") pod \"heat-api-8ff6c7648-wtgnm\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.445098 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-combined-ca-bundle\") pod \"heat-cfnapi-5d49fd5cb-kh9dz\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.445205 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-config-data-custom\") pod \"heat-api-8ff6c7648-wtgnm\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.445311 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-config-data\") pod \"heat-api-8ff6c7648-wtgnm\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.445475 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-combined-ca-bundle\") pod \"heat-api-8ff6c7648-wtgnm\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.445510 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-config-data\") pod \"heat-cfnapi-5d49fd5cb-kh9dz\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.451342 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-config-data-custom\") pod \"heat-cfnapi-5d49fd5cb-kh9dz\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.451641 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-combined-ca-bundle\") pod \"heat-api-8ff6c7648-wtgnm\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.454445 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-config-data\") pod \"heat-api-8ff6c7648-wtgnm\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.461414 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-config-data\") pod \"heat-cfnapi-5d49fd5cb-kh9dz\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.462636 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-config-data-custom\") pod \"heat-api-8ff6c7648-wtgnm\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.465189 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-combined-ca-bundle\") pod \"heat-cfnapi-5d49fd5cb-kh9dz\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.471144 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s47k6\" (UniqueName: \"kubernetes.io/projected/983d8aca-7ee4-4105-9678-63d608645975-kube-api-access-s47k6\") pod \"heat-api-8ff6c7648-wtgnm\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.480370 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlzgp\" (UniqueName: \"kubernetes.io/projected/ad488e84-ee1a-47f8-b626-fdc87919520d-kube-api-access-vlzgp\") pod \"heat-cfnapi-5d49fd5cb-kh9dz\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.524366 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.573839 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:01:59 crc kubenswrapper[4771]: I1002 10:01:59.592909 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:02:00 crc kubenswrapper[4771]: I1002 10:02:00.007309 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:02:00 crc kubenswrapper[4771]: I1002 10:02:00.015339 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-668c9ddd79-9f298" Oct 02 10:02:00 crc kubenswrapper[4771]: I1002 10:02:00.689558 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-svb6n" podUID="ec78b084-92a4-4eb9-ac22-bd7b51f4612c" containerName="registry-server" containerID="cri-o://4dd862c9c342a8db8e0fcaf6b368ca9bb165040c0c6a5f3c28edd84b62c30bdc" gracePeriod=2 Oct 02 10:02:00 crc kubenswrapper[4771]: I1002 10:02:00.710115 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:02:00 crc kubenswrapper[4771]: I1002 10:02:00.839560 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-hjvjj"] Oct 02 10:02:00 crc kubenswrapper[4771]: I1002 10:02:00.840492 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" podUID="452a7728-1a14-4649-b5bd-2ea19e39db4c" containerName="dnsmasq-dns" containerID="cri-o://9be0f1e2a3ff9da209f926f223a0fcd17605dd544224d89e0909b5bb670ab455" gracePeriod=10 Oct 02 10:02:01 crc kubenswrapper[4771]: I1002 10:02:01.743909 4771 generic.go:334] "Generic (PLEG): container finished" podID="452a7728-1a14-4649-b5bd-2ea19e39db4c" containerID="9be0f1e2a3ff9da209f926f223a0fcd17605dd544224d89e0909b5bb670ab455" exitCode=0 Oct 02 10:02:01 crc kubenswrapper[4771]: I1002 10:02:01.745881 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" event={"ID":"452a7728-1a14-4649-b5bd-2ea19e39db4c","Type":"ContainerDied","Data":"9be0f1e2a3ff9da209f926f223a0fcd17605dd544224d89e0909b5bb670ab455"} Oct 02 10:02:01 crc kubenswrapper[4771]: I1002 10:02:01.770684 4771 generic.go:334] "Generic (PLEG): container finished" podID="ec78b084-92a4-4eb9-ac22-bd7b51f4612c" containerID="4dd862c9c342a8db8e0fcaf6b368ca9bb165040c0c6a5f3c28edd84b62c30bdc" exitCode=0 Oct 02 10:02:01 crc kubenswrapper[4771]: I1002 10:02:01.770813 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svb6n" event={"ID":"ec78b084-92a4-4eb9-ac22-bd7b51f4612c","Type":"ContainerDied","Data":"4dd862c9c342a8db8e0fcaf6b368ca9bb165040c0c6a5f3c28edd84b62c30bdc"} Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.233795 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.393967 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-log-httpd\") pod \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.394117 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-config-data\") pod \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.394316 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5ms5\" (UniqueName: \"kubernetes.io/projected/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-kube-api-access-q5ms5\") pod \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.394366 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-sg-core-conf-yaml\") pod \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.394428 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-scripts\") pod \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.394493 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-run-httpd\") pod \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.394517 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-combined-ca-bundle\") pod \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\" (UID: \"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.395400 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" (UID: "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.397297 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" (UID: "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.420599 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-kube-api-access-q5ms5" (OuterVolumeSpecName: "kube-api-access-q5ms5") pod "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" (UID: "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f"). InnerVolumeSpecName "kube-api-access-q5ms5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.425586 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-scripts" (OuterVolumeSpecName: "scripts") pod "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" (UID: "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.503893 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" (UID: "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.505469 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5ms5\" (UniqueName: \"kubernetes.io/projected/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-kube-api-access-q5ms5\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.505515 4771 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.505528 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.505540 4771 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.505554 4771 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.535410 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.549472 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.745063 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-dns-swift-storage-0\") pod \"452a7728-1a14-4649-b5bd-2ea19e39db4c\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.745245 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-config\") pod \"452a7728-1a14-4649-b5bd-2ea19e39db4c\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.745365 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-ovsdbserver-nb\") pod \"452a7728-1a14-4649-b5bd-2ea19e39db4c\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.745408 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-catalog-content\") pod \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\" (UID: \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.745438 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-dns-svc\") pod \"452a7728-1a14-4649-b5bd-2ea19e39db4c\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.790748 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-utilities\") pod \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\" (UID: \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.790925 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zn4t\" (UniqueName: \"kubernetes.io/projected/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-kube-api-access-9zn4t\") pod \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\" (UID: \"ec78b084-92a4-4eb9-ac22-bd7b51f4612c\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.791081 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np2h9\" (UniqueName: \"kubernetes.io/projected/452a7728-1a14-4649-b5bd-2ea19e39db4c-kube-api-access-np2h9\") pod \"452a7728-1a14-4649-b5bd-2ea19e39db4c\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.791310 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-ovsdbserver-sb\") pod \"452a7728-1a14-4649-b5bd-2ea19e39db4c\" (UID: \"452a7728-1a14-4649-b5bd-2ea19e39db4c\") " Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.794293 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-utilities" (OuterVolumeSpecName: "utilities") pod "ec78b084-92a4-4eb9-ac22-bd7b51f4612c" (UID: "ec78b084-92a4-4eb9-ac22-bd7b51f4612c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:02 crc kubenswrapper[4771]: E1002 10:02:02.798112 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice/crio-8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f\": RecentStats: unable to find data in memory cache]" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.817475 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-kube-api-access-9zn4t" (OuterVolumeSpecName: "kube-api-access-9zn4t") pod "ec78b084-92a4-4eb9-ac22-bd7b51f4612c" (UID: "ec78b084-92a4-4eb9-ac22-bd7b51f4612c"). InnerVolumeSpecName "kube-api-access-9zn4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.825208 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/452a7728-1a14-4649-b5bd-2ea19e39db4c-kube-api-access-np2h9" (OuterVolumeSpecName: "kube-api-access-np2h9") pod "452a7728-1a14-4649-b5bd-2ea19e39db4c" (UID: "452a7728-1a14-4649-b5bd-2ea19e39db4c"). InnerVolumeSpecName "kube-api-access-np2h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.829159 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-config-data" (OuterVolumeSpecName: "config-data") pod "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" (UID: "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.904927 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.904969 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.904980 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zn4t\" (UniqueName: \"kubernetes.io/projected/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-kube-api-access-9zn4t\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.904994 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np2h9\" (UniqueName: \"kubernetes.io/projected/452a7728-1a14-4649-b5bd-2ea19e39db4c-kube-api-access-np2h9\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.918682 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "452a7728-1a14-4649-b5bd-2ea19e39db4c" (UID: "452a7728-1a14-4649-b5bd-2ea19e39db4c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.940150 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" event={"ID":"452a7728-1a14-4649-b5bd-2ea19e39db4c","Type":"ContainerDied","Data":"64d899e416a2b671f0fa00b71531d345ac16cc99372eb45159906bddce6247b0"} Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.940225 4771 scope.go:117] "RemoveContainer" containerID="9be0f1e2a3ff9da209f926f223a0fcd17605dd544224d89e0909b5bb670ab455" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.940386 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-hjvjj" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.986221 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "452a7728-1a14-4649-b5bd-2ea19e39db4c" (UID: "452a7728-1a14-4649-b5bd-2ea19e39db4c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:02:02 crc kubenswrapper[4771]: I1002 10:02:02.988146 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "452a7728-1a14-4649-b5bd-2ea19e39db4c" (UID: "452a7728-1a14-4649-b5bd-2ea19e39db4c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.002556 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cb800a24-4fb8-49e2-b5ee-7975a4f1f79f","Type":"ContainerDied","Data":"86964390f1dd2ff080457f91a091dc8c92b2f3352e7a275c41b0c493c9f150ca"} Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.002685 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.009295 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec78b084-92a4-4eb9-ac22-bd7b51f4612c" (UID: "ec78b084-92a4-4eb9-ac22-bd7b51f4612c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.012158 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.012200 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.012215 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec78b084-92a4-4eb9-ac22-bd7b51f4612c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.012229 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.048955 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-88f48768-bxk7h" event={"ID":"49ac894e-c59d-40e0-94e0-37c0913afd0f","Type":"ContainerStarted","Data":"8ea9c094796f03932a54d554459c3d27c03b97e50a099f9b6df39df8ad23ddc1"} Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.051684 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.061316 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" (UID: "cb800a24-4fb8-49e2-b5ee-7975a4f1f79f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:03 crc kubenswrapper[4771]: W1002 10:02:03.063377 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod983d8aca_7ee4_4105_9678_63d608645975.slice/crio-f37189e1dd8f9f61d420c868f671916ae3625cb4ac79e5bf9b52138076066217 WatchSource:0}: Error finding container f37189e1dd8f9f61d420c868f671916ae3625cb4ac79e5bf9b52138076066217: Status 404 returned error can't find the container with id f37189e1dd8f9f61d420c868f671916ae3625cb4ac79e5bf9b52138076066217 Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.063470 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-c7b979484-9sfpc" event={"ID":"9404916e-870c-403e-9320-25fda656bf6a","Type":"ContainerStarted","Data":"0aaa6967ff166fc2fc17c1686223d5544c03ca7c3464bd2da19e6c23163ee9f9"} Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.067078 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.096379 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svb6n" event={"ID":"ec78b084-92a4-4eb9-ac22-bd7b51f4612c","Type":"ContainerDied","Data":"976fba3bb30662f94c2388470dbee4ab0496bf29e0dbf1a6b38e0be5299e6841"} Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.096740 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svb6n" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.111645 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-8ff6c7648-wtgnm"] Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.132113 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.176859 4771 scope.go:117] "RemoveContainer" containerID="ea940924145b3c113e05d565e96d23b8813882966a0caa89309309a12e344864" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.177003 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5d49fd5cb-kh9dz"] Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.193739 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "452a7728-1a14-4649-b5bd-2ea19e39db4c" (UID: "452a7728-1a14-4649-b5bd-2ea19e39db4c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.197226 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-config" (OuterVolumeSpecName: "config") pod "452a7728-1a14-4649-b5bd-2ea19e39db4c" (UID: "452a7728-1a14-4649-b5bd-2ea19e39db4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.207395 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-88f48768-bxk7h" podStartSLOduration=5.21600409 podStartE2EDuration="13.207366198s" podCreationTimestamp="2025-10-02 10:01:50 +0000 UTC" firstStartedPulling="2025-10-02 10:01:53.042424634 +0000 UTC m=+1500.690109711" lastFinishedPulling="2025-10-02 10:02:01.033786752 +0000 UTC m=+1508.681471819" observedRunningTime="2025-10-02 10:02:03.074021414 +0000 UTC m=+1510.721706481" watchObservedRunningTime="2025-10-02 10:02:03.207366198 +0000 UTC m=+1510.855051275" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.207668 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-868b7bfb89-mjrmr"] Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.237098 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.237169 4771 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/452a7728-1a14-4649-b5bd-2ea19e39db4c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.242618 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-c7b979484-9sfpc" podStartSLOduration=4.808000313 podStartE2EDuration="13.242596278s" podCreationTimestamp="2025-10-02 10:01:50 +0000 UTC" firstStartedPulling="2025-10-02 10:01:52.540647549 +0000 UTC m=+1500.188332616" lastFinishedPulling="2025-10-02 10:02:00.975243514 +0000 UTC m=+1508.622928581" observedRunningTime="2025-10-02 10:02:03.11022049 +0000 UTC m=+1510.757905557" watchObservedRunningTime="2025-10-02 10:02:03.242596278 +0000 UTC m=+1510.890281345" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.608993 4771 scope.go:117] "RemoveContainer" containerID="bb7ea1b59d47aa056d3f4df5e6ca203ac5091bfe041d713a62b00549177eb312" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.712423 4771 scope.go:117] "RemoveContainer" containerID="48ffa7722cc5a44891b25a0bf45bdb35e6a21d55b848def01947b2101bd08f22" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.855283 4771 scope.go:117] "RemoveContainer" containerID="00e4574206a45d648bcda5196849131bf2c3fb95fbf0397513229cc9a789de07" Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.881072 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svb6n"] Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.894830 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-svb6n"] Oct 02 10:02:03 crc kubenswrapper[4771]: I1002 10:02:03.949059 4771 scope.go:117] "RemoveContainer" containerID="74bf1bb5db2b07de6f282f2c7a2c8ac004d12d5baae3d4580a58192f6649d863" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.023202 4771 scope.go:117] "RemoveContainer" containerID="4dd862c9c342a8db8e0fcaf6b368ca9bb165040c0c6a5f3c28edd84b62c30bdc" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.074265 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-hjvjj"] Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.091064 4771 scope.go:117] "RemoveContainer" containerID="ebc6196e862016bec9cda955179e9950d8ed43821f551ae7940259caa344ed3f" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.112204 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-hjvjj"] Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.166427 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.214908 4771 scope.go:117] "RemoveContainer" containerID="8849634ecc2b1267f585fa7418e9281eea2feb4defe0e84bcf5a6452a9d90d36" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.248173 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8ff6c7648-wtgnm" event={"ID":"983d8aca-7ee4-4105-9678-63d608645975","Type":"ContainerStarted","Data":"f37189e1dd8f9f61d420c868f671916ae3625cb4ac79e5bf9b52138076066217"} Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.250020 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.279520 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.295429 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:02:04 crc kubenswrapper[4771]: E1002 10:02:04.296118 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452a7728-1a14-4649-b5bd-2ea19e39db4c" containerName="init" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.296404 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="452a7728-1a14-4649-b5bd-2ea19e39db4c" containerName="init" Oct 02 10:02:04 crc kubenswrapper[4771]: E1002 10:02:04.296433 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="ceilometer-central-agent" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.296443 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="ceilometer-central-agent" Oct 02 10:02:04 crc kubenswrapper[4771]: E1002 10:02:04.296460 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec78b084-92a4-4eb9-ac22-bd7b51f4612c" containerName="extract-utilities" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.296471 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec78b084-92a4-4eb9-ac22-bd7b51f4612c" containerName="extract-utilities" Oct 02 10:02:04 crc kubenswrapper[4771]: E1002 10:02:04.296498 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec78b084-92a4-4eb9-ac22-bd7b51f4612c" containerName="registry-server" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.296507 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec78b084-92a4-4eb9-ac22-bd7b51f4612c" containerName="registry-server" Oct 02 10:02:04 crc kubenswrapper[4771]: E1002 10:02:04.296528 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="ceilometer-notification-agent" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.296539 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="ceilometer-notification-agent" Oct 02 10:02:04 crc kubenswrapper[4771]: E1002 10:02:04.296570 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="sg-core" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.296579 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="sg-core" Oct 02 10:02:04 crc kubenswrapper[4771]: E1002 10:02:04.296592 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="proxy-httpd" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.296602 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="proxy-httpd" Oct 02 10:02:04 crc kubenswrapper[4771]: E1002 10:02:04.296614 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec78b084-92a4-4eb9-ac22-bd7b51f4612c" containerName="extract-content" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.296623 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec78b084-92a4-4eb9-ac22-bd7b51f4612c" containerName="extract-content" Oct 02 10:02:04 crc kubenswrapper[4771]: E1002 10:02:04.296644 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452a7728-1a14-4649-b5bd-2ea19e39db4c" containerName="dnsmasq-dns" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.296652 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="452a7728-1a14-4649-b5bd-2ea19e39db4c" containerName="dnsmasq-dns" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.296950 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="ceilometer-central-agent" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.296971 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="proxy-httpd" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.296988 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec78b084-92a4-4eb9-ac22-bd7b51f4612c" containerName="registry-server" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.297002 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="ceilometer-notification-agent" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.297030 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" containerName="sg-core" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.297046 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="452a7728-1a14-4649-b5bd-2ea19e39db4c" containerName="dnsmasq-dns" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.300172 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.314751 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-config-data\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.314821 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e1a0457-04c3-4dd6-8065-a04d208525c9-run-httpd\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.314858 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.314897 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.315044 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e1a0457-04c3-4dd6-8065-a04d208525c9-log-httpd\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.320240 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-scripts\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.320556 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fnc5\" (UniqueName: \"kubernetes.io/projected/3e1a0457-04c3-4dd6-8065-a04d208525c9-kube-api-access-2fnc5\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.320855 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.321066 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.337624 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.341015 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" event={"ID":"ad488e84-ee1a-47f8-b626-fdc87919520d","Type":"ContainerStarted","Data":"36d8e029971d773d213042f48869ced689910f3bfceb01d6a62d43f0af2b7b6e"} Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.351110 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-8ff6c7648-wtgnm" podStartSLOduration=5.351079106 podStartE2EDuration="5.351079106s" podCreationTimestamp="2025-10-02 10:01:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:02:04.32600136 +0000 UTC m=+1511.973686427" watchObservedRunningTime="2025-10-02 10:02:04.351079106 +0000 UTC m=+1511.998764173" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.351368 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.371214 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-868b7bfb89-mjrmr" event={"ID":"a4bc89da-2332-412a-b7af-6056f830310a","Type":"ContainerStarted","Data":"c99213ee773c091ec2ec06c5e15dc68d76582db6e7aa50cea17319754bdfb116"} Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.372364 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.423829 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-scripts\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.423914 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fnc5\" (UniqueName: \"kubernetes.io/projected/3e1a0457-04c3-4dd6-8065-a04d208525c9-kube-api-access-2fnc5\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.424036 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-config-data\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.424059 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e1a0457-04c3-4dd6-8065-a04d208525c9-run-httpd\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.424084 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.424118 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.425803 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e1a0457-04c3-4dd6-8065-a04d208525c9-log-httpd\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.427195 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e1a0457-04c3-4dd6-8065-a04d208525c9-log-httpd\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.428430 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e1a0457-04c3-4dd6-8065-a04d208525c9-run-httpd\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.428869 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-868b7bfb89-mjrmr" podStartSLOduration=5.428841912 podStartE2EDuration="5.428841912s" podCreationTimestamp="2025-10-02 10:01:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:02:04.392434631 +0000 UTC m=+1512.040119708" watchObservedRunningTime="2025-10-02 10:02:04.428841912 +0000 UTC m=+1512.076526979" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.442692 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" podStartSLOduration=5.442671015 podStartE2EDuration="5.442671015s" podCreationTimestamp="2025-10-02 10:01:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:02:04.411046822 +0000 UTC m=+1512.058731899" watchObservedRunningTime="2025-10-02 10:02:04.442671015 +0000 UTC m=+1512.090356082" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.451973 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fnc5\" (UniqueName: \"kubernetes.io/projected/3e1a0457-04c3-4dd6-8065-a04d208525c9-kube-api-access-2fnc5\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.452268 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.452500 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.452786 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-scripts\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.465776 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-config-data\") pod \"ceilometer-0\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " pod="openstack/ceilometer-0" Oct 02 10:02:04 crc kubenswrapper[4771]: I1002 10:02:04.712959 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.197073 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-88f48768-bxk7h"] Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.311027 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-c7b979484-9sfpc"] Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.366283 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6dfdb97779-bx9w5"] Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.368106 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.403300 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.403523 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5cc7b668b9-5jkkb"] Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.403662 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.413651 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.418228 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.418450 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.532398 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6dfdb97779-bx9w5"] Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.533964 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-internal-tls-certs\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.534012 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-combined-ca-bundle\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.534062 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4t5c\" (UniqueName: \"kubernetes.io/projected/f740f4c4-0587-4077-8a4c-3135aea31447-kube-api-access-d4t5c\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.534171 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-config-data-custom\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.534251 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-config-data\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.534286 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-config-data\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.534322 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vc74\" (UniqueName: \"kubernetes.io/projected/ff726ea6-69e3-4d3e-98b5-099265579c80-kube-api-access-5vc74\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.534407 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-internal-tls-certs\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.534441 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-combined-ca-bundle\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.534461 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-public-tls-certs\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.534497 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-public-tls-certs\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.534521 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-config-data-custom\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.546090 4771 generic.go:334] "Generic (PLEG): container finished" podID="983d8aca-7ee4-4105-9678-63d608645975" containerID="7da5c167f5ae160c62693fabf95dfa33796b0c77b3230008807dbeccc0ef38e5" exitCode=1 Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.546411 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8ff6c7648-wtgnm" event={"ID":"983d8aca-7ee4-4105-9678-63d608645975","Type":"ContainerDied","Data":"7da5c167f5ae160c62693fabf95dfa33796b0c77b3230008807dbeccc0ef38e5"} Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.547238 4771 scope.go:117] "RemoveContainer" containerID="7da5c167f5ae160c62693fabf95dfa33796b0c77b3230008807dbeccc0ef38e5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.588090 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5cc7b668b9-5jkkb"] Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.600052 4771 generic.go:334] "Generic (PLEG): container finished" podID="ad488e84-ee1a-47f8-b626-fdc87919520d" containerID="338cee197d74017bc57c09da1f4f16c5947acdfb660658eb7ea2005a6cf2a35c" exitCode=1 Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.600147 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" event={"ID":"ad488e84-ee1a-47f8-b626-fdc87919520d","Type":"ContainerDied","Data":"338cee197d74017bc57c09da1f4f16c5947acdfb660658eb7ea2005a6cf2a35c"} Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.601578 4771 scope.go:117] "RemoveContainer" containerID="338cee197d74017bc57c09da1f4f16c5947acdfb660658eb7ea2005a6cf2a35c" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.612166 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-88f48768-bxk7h" podUID="49ac894e-c59d-40e0-94e0-37c0913afd0f" containerName="heat-api" containerID="cri-o://8ea9c094796f03932a54d554459c3d27c03b97e50a099f9b6df39df8ad23ddc1" gracePeriod=60 Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.613790 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-868b7bfb89-mjrmr" event={"ID":"a4bc89da-2332-412a-b7af-6056f830310a","Type":"ContainerStarted","Data":"763b93cb9e0bd45f8673736c7da2c1d7ebc5601f38a1b419d4f03d0dea0ff226"} Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.613983 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-c7b979484-9sfpc" podUID="9404916e-870c-403e-9320-25fda656bf6a" containerName="heat-cfnapi" containerID="cri-o://0aaa6967ff166fc2fc17c1686223d5544c03ca7c3464bd2da19e6c23163ee9f9" gracePeriod=60 Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.637503 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-config-data\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.637586 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-config-data\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.637630 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vc74\" (UniqueName: \"kubernetes.io/projected/ff726ea6-69e3-4d3e-98b5-099265579c80-kube-api-access-5vc74\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.637771 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-internal-tls-certs\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.637825 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-combined-ca-bundle\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.637842 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-public-tls-certs\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.637913 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-public-tls-certs\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.637947 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-config-data-custom\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.638069 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-internal-tls-certs\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.638102 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-combined-ca-bundle\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.638252 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4t5c\" (UniqueName: \"kubernetes.io/projected/f740f4c4-0587-4077-8a4c-3135aea31447-kube-api-access-d4t5c\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.638282 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-config-data-custom\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.650177 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-config-data-custom\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.653536 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-config-data\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.656337 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-public-tls-certs\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.656923 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-internal-tls-certs\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.657691 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-config-data\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.658049 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vc74\" (UniqueName: \"kubernetes.io/projected/ff726ea6-69e3-4d3e-98b5-099265579c80-kube-api-access-5vc74\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.661092 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-combined-ca-bundle\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.663445 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-public-tls-certs\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.663692 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-combined-ca-bundle\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.669051 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4t5c\" (UniqueName: \"kubernetes.io/projected/f740f4c4-0587-4077-8a4c-3135aea31447-kube-api-access-d4t5c\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.703651 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-config-data-custom\") pod \"heat-api-6dfdb97779-bx9w5\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.706871 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-internal-tls-certs\") pod \"heat-cfnapi-5cc7b668b9-5jkkb\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.774706 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="452a7728-1a14-4649-b5bd-2ea19e39db4c" path="/var/lib/kubelet/pods/452a7728-1a14-4649-b5bd-2ea19e39db4c/volumes" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.779568 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.798251 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb800a24-4fb8-49e2-b5ee-7975a4f1f79f" path="/var/lib/kubelet/pods/cb800a24-4fb8-49e2-b5ee-7975a4f1f79f/volumes" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.861485 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec78b084-92a4-4eb9-ac22-bd7b51f4612c" path="/var/lib/kubelet/pods/ec78b084-92a4-4eb9-ac22-bd7b51f4612c/volumes" Oct 02 10:02:05 crc kubenswrapper[4771]: I1002 10:02:05.863630 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:06 crc kubenswrapper[4771]: I1002 10:02:06.144052 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:02:06 crc kubenswrapper[4771]: I1002 10:02:06.643250 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8ff6c7648-wtgnm" event={"ID":"983d8aca-7ee4-4105-9678-63d608645975","Type":"ContainerStarted","Data":"f3bd35ae1d18f4913d4344dba97cb1172d9600dd4b838f23985dd92ef30a1368"} Oct 02 10:02:06 crc kubenswrapper[4771]: I1002 10:02:06.643719 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:02:06 crc kubenswrapper[4771]: I1002 10:02:06.659512 4771 generic.go:334] "Generic (PLEG): container finished" podID="49ac894e-c59d-40e0-94e0-37c0913afd0f" containerID="8ea9c094796f03932a54d554459c3d27c03b97e50a099f9b6df39df8ad23ddc1" exitCode=0 Oct 02 10:02:06 crc kubenswrapper[4771]: I1002 10:02:06.659814 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-88f48768-bxk7h" event={"ID":"49ac894e-c59d-40e0-94e0-37c0913afd0f","Type":"ContainerDied","Data":"8ea9c094796f03932a54d554459c3d27c03b97e50a099f9b6df39df8ad23ddc1"} Oct 02 10:02:06 crc kubenswrapper[4771]: I1002 10:02:06.675583 4771 generic.go:334] "Generic (PLEG): container finished" podID="9404916e-870c-403e-9320-25fda656bf6a" containerID="0aaa6967ff166fc2fc17c1686223d5544c03ca7c3464bd2da19e6c23163ee9f9" exitCode=0 Oct 02 10:02:06 crc kubenswrapper[4771]: I1002 10:02:06.675712 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-c7b979484-9sfpc" event={"ID":"9404916e-870c-403e-9320-25fda656bf6a","Type":"ContainerDied","Data":"0aaa6967ff166fc2fc17c1686223d5544c03ca7c3464bd2da19e6c23163ee9f9"} Oct 02 10:02:06 crc kubenswrapper[4771]: I1002 10:02:06.693285 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e1a0457-04c3-4dd6-8065-a04d208525c9","Type":"ContainerStarted","Data":"d376784f90450d7039ab9198a34157220abc1f9e433a482d0d295104a844b3e9"} Oct 02 10:02:06 crc kubenswrapper[4771]: I1002 10:02:06.705013 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" event={"ID":"ad488e84-ee1a-47f8-b626-fdc87919520d","Type":"ContainerStarted","Data":"63a6291c1d4af868fb7e8bd6b0a8237a180876c54733b17a026cb296cbe0ce71"} Oct 02 10:02:06 crc kubenswrapper[4771]: I1002 10:02:06.705096 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:02:07 crc kubenswrapper[4771]: E1002 10:02:07.238081 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice/crio-8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice\": RecentStats: unable to find data in memory cache]" Oct 02 10:02:07 crc kubenswrapper[4771]: I1002 10:02:07.280853 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5cc7b668b9-5jkkb"] Oct 02 10:02:07 crc kubenswrapper[4771]: I1002 10:02:07.375765 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6dfdb97779-bx9w5"] Oct 02 10:02:07 crc kubenswrapper[4771]: I1002 10:02:07.721640 4771 generic.go:334] "Generic (PLEG): container finished" podID="983d8aca-7ee4-4105-9678-63d608645975" containerID="f3bd35ae1d18f4913d4344dba97cb1172d9600dd4b838f23985dd92ef30a1368" exitCode=1 Oct 02 10:02:07 crc kubenswrapper[4771]: I1002 10:02:07.721724 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8ff6c7648-wtgnm" event={"ID":"983d8aca-7ee4-4105-9678-63d608645975","Type":"ContainerDied","Data":"f3bd35ae1d18f4913d4344dba97cb1172d9600dd4b838f23985dd92ef30a1368"} Oct 02 10:02:07 crc kubenswrapper[4771]: I1002 10:02:07.721820 4771 scope.go:117] "RemoveContainer" containerID="7da5c167f5ae160c62693fabf95dfa33796b0c77b3230008807dbeccc0ef38e5" Oct 02 10:02:07 crc kubenswrapper[4771]: I1002 10:02:07.722506 4771 scope.go:117] "RemoveContainer" containerID="f3bd35ae1d18f4913d4344dba97cb1172d9600dd4b838f23985dd92ef30a1368" Oct 02 10:02:07 crc kubenswrapper[4771]: E1002 10:02:07.722904 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-8ff6c7648-wtgnm_openstack(983d8aca-7ee4-4105-9678-63d608645975)\"" pod="openstack/heat-api-8ff6c7648-wtgnm" podUID="983d8aca-7ee4-4105-9678-63d608645975" Oct 02 10:02:07 crc kubenswrapper[4771]: I1002 10:02:07.726384 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e1a0457-04c3-4dd6-8065-a04d208525c9","Type":"ContainerStarted","Data":"fbe281d1e373fb76203172136537c17314552d8df3d2fa6c50de186d8b9cbd15"} Oct 02 10:02:07 crc kubenswrapper[4771]: I1002 10:02:07.733064 4771 generic.go:334] "Generic (PLEG): container finished" podID="ad488e84-ee1a-47f8-b626-fdc87919520d" containerID="63a6291c1d4af868fb7e8bd6b0a8237a180876c54733b17a026cb296cbe0ce71" exitCode=1 Oct 02 10:02:07 crc kubenswrapper[4771]: I1002 10:02:07.733229 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" event={"ID":"ad488e84-ee1a-47f8-b626-fdc87919520d","Type":"ContainerDied","Data":"63a6291c1d4af868fb7e8bd6b0a8237a180876c54733b17a026cb296cbe0ce71"} Oct 02 10:02:07 crc kubenswrapper[4771]: I1002 10:02:07.734003 4771 scope.go:117] "RemoveContainer" containerID="63a6291c1d4af868fb7e8bd6b0a8237a180876c54733b17a026cb296cbe0ce71" Oct 02 10:02:07 crc kubenswrapper[4771]: E1002 10:02:07.734326 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-5d49fd5cb-kh9dz_openstack(ad488e84-ee1a-47f8-b626-fdc87919520d)\"" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" Oct 02 10:02:08 crc kubenswrapper[4771]: I1002 10:02:08.752885 4771 scope.go:117] "RemoveContainer" containerID="63a6291c1d4af868fb7e8bd6b0a8237a180876c54733b17a026cb296cbe0ce71" Oct 02 10:02:08 crc kubenswrapper[4771]: I1002 10:02:08.753483 4771 scope.go:117] "RemoveContainer" containerID="f3bd35ae1d18f4913d4344dba97cb1172d9600dd4b838f23985dd92ef30a1368" Oct 02 10:02:08 crc kubenswrapper[4771]: E1002 10:02:08.753507 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-5d49fd5cb-kh9dz_openstack(ad488e84-ee1a-47f8-b626-fdc87919520d)\"" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" Oct 02 10:02:08 crc kubenswrapper[4771]: E1002 10:02:08.755903 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-8ff6c7648-wtgnm_openstack(983d8aca-7ee4-4105-9678-63d608645975)\"" pod="openstack/heat-api-8ff6c7648-wtgnm" podUID="983d8aca-7ee4-4105-9678-63d608645975" Oct 02 10:02:09 crc kubenswrapper[4771]: I1002 10:02:09.575147 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:02:09 crc kubenswrapper[4771]: I1002 10:02:09.593858 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:02:09 crc kubenswrapper[4771]: I1002 10:02:09.763944 4771 scope.go:117] "RemoveContainer" containerID="63a6291c1d4af868fb7e8bd6b0a8237a180876c54733b17a026cb296cbe0ce71" Oct 02 10:02:09 crc kubenswrapper[4771]: I1002 10:02:09.764034 4771 scope.go:117] "RemoveContainer" containerID="f3bd35ae1d18f4913d4344dba97cb1172d9600dd4b838f23985dd92ef30a1368" Oct 02 10:02:09 crc kubenswrapper[4771]: E1002 10:02:09.764276 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-5d49fd5cb-kh9dz_openstack(ad488e84-ee1a-47f8-b626-fdc87919520d)\"" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" Oct 02 10:02:09 crc kubenswrapper[4771]: E1002 10:02:09.764396 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-8ff6c7648-wtgnm_openstack(983d8aca-7ee4-4105-9678-63d608645975)\"" pod="openstack/heat-api-8ff6c7648-wtgnm" podUID="983d8aca-7ee4-4105-9678-63d608645975" Oct 02 10:02:10 crc kubenswrapper[4771]: I1002 10:02:10.392898 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:02:11 crc kubenswrapper[4771]: I1002 10:02:11.151229 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:02:12 crc kubenswrapper[4771]: I1002 10:02:12.146161 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:02:12 crc kubenswrapper[4771]: I1002 10:02:12.146244 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:02:14 crc kubenswrapper[4771]: I1002 10:02:14.856088 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-66d599bfb9-7qmxw" Oct 02 10:02:14 crc kubenswrapper[4771]: I1002 10:02:14.936499 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7d8b77dfb-5n9ct"] Oct 02 10:02:14 crc kubenswrapper[4771]: I1002 10:02:14.936865 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7d8b77dfb-5n9ct" podUID="8f051c06-1d08-41da-9c70-8769a08d3fa1" containerName="neutron-api" containerID="cri-o://375d1f6c5d380d68ddb81d6180f95a3549ace994c00a2e19f4e9ab1801cc148a" gracePeriod=30 Oct 02 10:02:14 crc kubenswrapper[4771]: I1002 10:02:14.937431 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7d8b77dfb-5n9ct" podUID="8f051c06-1d08-41da-9c70-8769a08d3fa1" containerName="neutron-httpd" containerID="cri-o://1a83ca34e9be89357ddfe9811bc59a269810259a0885c84ee55ea230989d8975" gracePeriod=30 Oct 02 10:02:15 crc kubenswrapper[4771]: W1002 10:02:15.377346 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf740f4c4_0587_4077_8a4c_3135aea31447.slice/crio-1d6a04c992549b1ed393df3a48c5dbc95049e26272552539b71078e6a6fe2f86 WatchSource:0}: Error finding container 1d6a04c992549b1ed393df3a48c5dbc95049e26272552539b71078e6a6fe2f86: Status 404 returned error can't find the container with id 1d6a04c992549b1ed393df3a48c5dbc95049e26272552539b71078e6a6fe2f86 Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.677261 4771 scope.go:117] "RemoveContainer" containerID="338cee197d74017bc57c09da1f4f16c5947acdfb660658eb7ea2005a6cf2a35c" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.784406 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.815446 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.823244 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsvc2\" (UniqueName: \"kubernetes.io/projected/49ac894e-c59d-40e0-94e0-37c0913afd0f-kube-api-access-wsvc2\") pod \"49ac894e-c59d-40e0-94e0-37c0913afd0f\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.823395 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-combined-ca-bundle\") pod \"49ac894e-c59d-40e0-94e0-37c0913afd0f\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.823510 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-config-data\") pod \"49ac894e-c59d-40e0-94e0-37c0913afd0f\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.831420 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-config-data-custom\") pod \"49ac894e-c59d-40e0-94e0-37c0913afd0f\" (UID: \"49ac894e-c59d-40e0-94e0-37c0913afd0f\") " Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.855669 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ac894e-c59d-40e0-94e0-37c0913afd0f-kube-api-access-wsvc2" (OuterVolumeSpecName: "kube-api-access-wsvc2") pod "49ac894e-c59d-40e0-94e0-37c0913afd0f" (UID: "49ac894e-c59d-40e0-94e0-37c0913afd0f"). InnerVolumeSpecName "kube-api-access-wsvc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.857453 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "49ac894e-c59d-40e0-94e0-37c0913afd0f" (UID: "49ac894e-c59d-40e0-94e0-37c0913afd0f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.884918 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" event={"ID":"f740f4c4-0587-4077-8a4c-3135aea31447","Type":"ContainerStarted","Data":"1d6a04c992549b1ed393df3a48c5dbc95049e26272552539b71078e6a6fe2f86"} Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.886789 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6dfdb97779-bx9w5" event={"ID":"ff726ea6-69e3-4d3e-98b5-099265579c80","Type":"ContainerStarted","Data":"f05783dcc9d8e306cabb96eb5e48ec0cb5208ef7f0f23366fbe3b5a63afa9ab6"} Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.888492 4771 generic.go:334] "Generic (PLEG): container finished" podID="8f051c06-1d08-41da-9c70-8769a08d3fa1" containerID="1a83ca34e9be89357ddfe9811bc59a269810259a0885c84ee55ea230989d8975" exitCode=0 Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.888584 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d8b77dfb-5n9ct" event={"ID":"8f051c06-1d08-41da-9c70-8769a08d3fa1","Type":"ContainerDied","Data":"1a83ca34e9be89357ddfe9811bc59a269810259a0885c84ee55ea230989d8975"} Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.912320 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-88f48768-bxk7h" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.913344 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-88f48768-bxk7h" event={"ID":"49ac894e-c59d-40e0-94e0-37c0913afd0f","Type":"ContainerDied","Data":"556b37d3633a6fa229aaaf1ef92b3ca46075ae7bbd04e687b790fa8b0beaeb7d"} Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.913398 4771 scope.go:117] "RemoveContainer" containerID="8ea9c094796f03932a54d554459c3d27c03b97e50a099f9b6df39df8ad23ddc1" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.921235 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-c7b979484-9sfpc" event={"ID":"9404916e-870c-403e-9320-25fda656bf6a","Type":"ContainerDied","Data":"28916e258dc641e41144e4e82834d76077acaa926d2e92258a6a4e294d834abf"} Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.921339 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-c7b979484-9sfpc" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.933255 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49ac894e-c59d-40e0-94e0-37c0913afd0f" (UID: "49ac894e-c59d-40e0-94e0-37c0913afd0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.934611 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-config-data\") pod \"9404916e-870c-403e-9320-25fda656bf6a\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.935421 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-combined-ca-bundle\") pod \"9404916e-870c-403e-9320-25fda656bf6a\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.935618 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-config-data-custom\") pod \"9404916e-870c-403e-9320-25fda656bf6a\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.935724 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmq68\" (UniqueName: \"kubernetes.io/projected/9404916e-870c-403e-9320-25fda656bf6a-kube-api-access-wmq68\") pod \"9404916e-870c-403e-9320-25fda656bf6a\" (UID: \"9404916e-870c-403e-9320-25fda656bf6a\") " Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.936548 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.936713 4771 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.936810 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsvc2\" (UniqueName: \"kubernetes.io/projected/49ac894e-c59d-40e0-94e0-37c0913afd0f-kube-api-access-wsvc2\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.962836 4771 scope.go:117] "RemoveContainer" containerID="0aaa6967ff166fc2fc17c1686223d5544c03ca7c3464bd2da19e6c23163ee9f9" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.963582 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9404916e-870c-403e-9320-25fda656bf6a-kube-api-access-wmq68" (OuterVolumeSpecName: "kube-api-access-wmq68") pod "9404916e-870c-403e-9320-25fda656bf6a" (UID: "9404916e-870c-403e-9320-25fda656bf6a"). InnerVolumeSpecName "kube-api-access-wmq68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:15 crc kubenswrapper[4771]: I1002 10:02:15.965377 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9404916e-870c-403e-9320-25fda656bf6a" (UID: "9404916e-870c-403e-9320-25fda656bf6a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.021433 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9404916e-870c-403e-9320-25fda656bf6a" (UID: "9404916e-870c-403e-9320-25fda656bf6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.039228 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.039271 4771 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.039281 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmq68\" (UniqueName: \"kubernetes.io/projected/9404916e-870c-403e-9320-25fda656bf6a-kube-api-access-wmq68\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.040291 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-config-data" (OuterVolumeSpecName: "config-data") pod "49ac894e-c59d-40e0-94e0-37c0913afd0f" (UID: "49ac894e-c59d-40e0-94e0-37c0913afd0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.076495 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-config-data" (OuterVolumeSpecName: "config-data") pod "9404916e-870c-403e-9320-25fda656bf6a" (UID: "9404916e-870c-403e-9320-25fda656bf6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.141420 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9404916e-870c-403e-9320-25fda656bf6a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.141464 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49ac894e-c59d-40e0-94e0-37c0913afd0f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.261791 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-88f48768-bxk7h"] Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.275379 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-88f48768-bxk7h"] Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.291560 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-c7b979484-9sfpc"] Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.303396 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-c7b979484-9sfpc"] Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.937567 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.939179 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.939295 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6dfdb97779-bx9w5" event={"ID":"ff726ea6-69e3-4d3e-98b5-099265579c80","Type":"ContainerStarted","Data":"aaa1d2cffcab599d96ef4dc30e9419a93554d739b80468d87c9266d4999d5f94"} Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.939820 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="69624004-f6bf-48f2-af01-44b450260c6a" containerName="glance-log" containerID="cri-o://5d9dabe27276c37383ba954078c45957a44316629dc09a6b9c7f3b7d8cb06fa8" gracePeriod=30 Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.939820 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="69624004-f6bf-48f2-af01-44b450260c6a" containerName="glance-httpd" containerID="cri-o://8a93dc16801ca30b050caf35b39ec4c79462b62b0bc15228517efa1865e06820" gracePeriod=30 Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.951188 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" event={"ID":"f740f4c4-0587-4077-8a4c-3135aea31447","Type":"ContainerStarted","Data":"05e51e74039fab8e39bc3c070ba2da03f715ddff25ed226df376e4cc45e96f3f"} Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.953033 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.964392 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6dfdb97779-bx9w5" podStartSLOduration=11.964366485 podStartE2EDuration="11.964366485s" podCreationTimestamp="2025-10-02 10:02:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:02:16.960583013 +0000 UTC m=+1524.608268100" watchObservedRunningTime="2025-10-02 10:02:16.964366485 +0000 UTC m=+1524.612051572" Oct 02 10:02:16 crc kubenswrapper[4771]: I1002 10:02:16.992755 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" podStartSLOduration=11.99273418 podStartE2EDuration="11.99273418s" podCreationTimestamp="2025-10-02 10:02:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:02:16.983943583 +0000 UTC m=+1524.631628660" watchObservedRunningTime="2025-10-02 10:02:16.99273418 +0000 UTC m=+1524.640419247" Oct 02 10:02:17 crc kubenswrapper[4771]: E1002 10:02:17.543945 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice/crio-8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f\": RecentStats: unable to find data in memory cache]" Oct 02 10:02:17 crc kubenswrapper[4771]: E1002 10:02:17.545669 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice/crio-8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice\": RecentStats: unable to find data in memory cache]" Oct 02 10:02:17 crc kubenswrapper[4771]: I1002 10:02:17.700156 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ac894e-c59d-40e0-94e0-37c0913afd0f" path="/var/lib/kubelet/pods/49ac894e-c59d-40e0-94e0-37c0913afd0f/volumes" Oct 02 10:02:17 crc kubenswrapper[4771]: I1002 10:02:17.703712 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9404916e-870c-403e-9320-25fda656bf6a" path="/var/lib/kubelet/pods/9404916e-870c-403e-9320-25fda656bf6a/volumes" Oct 02 10:02:17 crc kubenswrapper[4771]: I1002 10:02:17.986710 4771 generic.go:334] "Generic (PLEG): container finished" podID="69624004-f6bf-48f2-af01-44b450260c6a" containerID="5d9dabe27276c37383ba954078c45957a44316629dc09a6b9c7f3b7d8cb06fa8" exitCode=143 Oct 02 10:02:17 crc kubenswrapper[4771]: I1002 10:02:17.988595 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"69624004-f6bf-48f2-af01-44b450260c6a","Type":"ContainerDied","Data":"5d9dabe27276c37383ba954078c45957a44316629dc09a6b9c7f3b7d8cb06fa8"} Oct 02 10:02:18 crc kubenswrapper[4771]: I1002 10:02:18.580216 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:02:18 crc kubenswrapper[4771]: I1002 10:02:18.580870 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4ecc1a24-b566-4f4b-8a39-10b49533cd66" containerName="glance-log" containerID="cri-o://25ace71c32266c58afbd34ba55f7b5a6a6382d12779760e38f072b19f23cb919" gracePeriod=30 Oct 02 10:02:18 crc kubenswrapper[4771]: I1002 10:02:18.580937 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4ecc1a24-b566-4f4b-8a39-10b49533cd66" containerName="glance-httpd" containerID="cri-o://206ed65503085f9fbdac82b37985708060c35a2884d9991a7324b4d264d3228f" gracePeriod=30 Oct 02 10:02:19 crc kubenswrapper[4771]: I1002 10:02:19.000610 4771 generic.go:334] "Generic (PLEG): container finished" podID="4ecc1a24-b566-4f4b-8a39-10b49533cd66" containerID="25ace71c32266c58afbd34ba55f7b5a6a6382d12779760e38f072b19f23cb919" exitCode=143 Oct 02 10:02:19 crc kubenswrapper[4771]: I1002 10:02:19.000826 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4ecc1a24-b566-4f4b-8a39-10b49533cd66","Type":"ContainerDied","Data":"25ace71c32266c58afbd34ba55f7b5a6a6382d12779760e38f072b19f23cb919"} Oct 02 10:02:19 crc kubenswrapper[4771]: I1002 10:02:19.003381 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e1a0457-04c3-4dd6-8065-a04d208525c9","Type":"ContainerStarted","Data":"a3912405d27c0f84507547d3e57a9a434827c85fe0ac745c017c03df7dba93fb"} Oct 02 10:02:19 crc kubenswrapper[4771]: I1002 10:02:19.006791 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"91134808-2774-4b54-8a59-09e9447dd87f","Type":"ContainerStarted","Data":"f0e4c3eb9fae43787e74606f47de39e7e0b967a85415081eadfe1a5b455cc347"} Oct 02 10:02:19 crc kubenswrapper[4771]: I1002 10:02:19.027515 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.472610142 podStartE2EDuration="41.027476655s" podCreationTimestamp="2025-10-02 10:01:38 +0000 UTC" firstStartedPulling="2025-10-02 10:01:40.203182755 +0000 UTC m=+1487.850867822" lastFinishedPulling="2025-10-02 10:02:17.758049268 +0000 UTC m=+1525.405734335" observedRunningTime="2025-10-02 10:02:19.021708159 +0000 UTC m=+1526.669393236" watchObservedRunningTime="2025-10-02 10:02:19.027476655 +0000 UTC m=+1526.675161722" Oct 02 10:02:19 crc kubenswrapper[4771]: I1002 10:02:19.594205 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:02:19 crc kubenswrapper[4771]: I1002 10:02:19.652834 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-649769f8cf-k8cph"] Oct 02 10:02:19 crc kubenswrapper[4771]: I1002 10:02:19.653066 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-649769f8cf-k8cph" podUID="d7b40340-a806-490f-9842-a074ffba2e93" containerName="heat-engine" containerID="cri-o://d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98" gracePeriod=60 Oct 02 10:02:20 crc kubenswrapper[4771]: I1002 10:02:20.205462 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:02:20 crc kubenswrapper[4771]: E1002 10:02:20.340029 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 02 10:02:20 crc kubenswrapper[4771]: E1002 10:02:20.341735 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 02 10:02:20 crc kubenswrapper[4771]: E1002 10:02:20.344113 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 02 10:02:20 crc kubenswrapper[4771]: E1002 10:02:20.344194 4771 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-649769f8cf-k8cph" podUID="d7b40340-a806-490f-9842-a074ffba2e93" containerName="heat-engine" Oct 02 10:02:20 crc kubenswrapper[4771]: I1002 10:02:20.798649 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-c7b979484-9sfpc" podUID="9404916e-870c-403e-9320-25fda656bf6a" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.216:8000/healthcheck\": dial tcp 10.217.0.216:8000: i/o timeout (Client.Timeout exceeded while awaiting headers)" Oct 02 10:02:20 crc kubenswrapper[4771]: I1002 10:02:20.852251 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-88f48768-bxk7h" podUID="49ac894e-c59d-40e0-94e0-37c0913afd0f" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.217:8004/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.043533 4771 generic.go:334] "Generic (PLEG): container finished" podID="8f051c06-1d08-41da-9c70-8769a08d3fa1" containerID="375d1f6c5d380d68ddb81d6180f95a3549ace994c00a2e19f4e9ab1801cc148a" exitCode=0 Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.043587 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d8b77dfb-5n9ct" event={"ID":"8f051c06-1d08-41da-9c70-8769a08d3fa1","Type":"ContainerDied","Data":"375d1f6c5d380d68ddb81d6180f95a3549ace994c00a2e19f4e9ab1801cc148a"} Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.046572 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e1a0457-04c3-4dd6-8065-a04d208525c9","Type":"ContainerStarted","Data":"aebc0b89a5d5628ec59f5be0429060dd4ca06f624db3f883833edf6a3e93229b"} Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.050068 4771 generic.go:334] "Generic (PLEG): container finished" podID="69624004-f6bf-48f2-af01-44b450260c6a" containerID="8a93dc16801ca30b050caf35b39ec4c79462b62b0bc15228517efa1865e06820" exitCode=0 Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.050106 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"69624004-f6bf-48f2-af01-44b450260c6a","Type":"ContainerDied","Data":"8a93dc16801ca30b050caf35b39ec4c79462b62b0bc15228517efa1865e06820"} Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.331590 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-fhkn7"] Oct 02 10:02:21 crc kubenswrapper[4771]: E1002 10:02:21.332394 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49ac894e-c59d-40e0-94e0-37c0913afd0f" containerName="heat-api" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.332412 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="49ac894e-c59d-40e0-94e0-37c0913afd0f" containerName="heat-api" Oct 02 10:02:21 crc kubenswrapper[4771]: E1002 10:02:21.332509 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9404916e-870c-403e-9320-25fda656bf6a" containerName="heat-cfnapi" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.332519 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9404916e-870c-403e-9320-25fda656bf6a" containerName="heat-cfnapi" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.333001 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="9404916e-870c-403e-9320-25fda656bf6a" containerName="heat-cfnapi" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.333026 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="49ac894e-c59d-40e0-94e0-37c0913afd0f" containerName="heat-api" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.334167 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-fhkn7" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.358023 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-fhkn7"] Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.358637 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.431946 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvn7n\" (UniqueName: \"kubernetes.io/projected/a6e71c3b-c209-4585-a6d4-e951dbf21b42-kube-api-access-dvn7n\") pod \"nova-api-db-create-fhkn7\" (UID: \"a6e71c3b-c209-4585-a6d4-e951dbf21b42\") " pod="openstack/nova-api-db-create-fhkn7" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.509818 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-lpdl4"] Oct 02 10:02:21 crc kubenswrapper[4771]: E1002 10:02:21.511485 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f051c06-1d08-41da-9c70-8769a08d3fa1" containerName="neutron-httpd" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.511513 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f051c06-1d08-41da-9c70-8769a08d3fa1" containerName="neutron-httpd" Oct 02 10:02:21 crc kubenswrapper[4771]: E1002 10:02:21.511648 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f051c06-1d08-41da-9c70-8769a08d3fa1" containerName="neutron-api" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.511658 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f051c06-1d08-41da-9c70-8769a08d3fa1" containerName="neutron-api" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.511962 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f051c06-1d08-41da-9c70-8769a08d3fa1" containerName="neutron-httpd" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.512007 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f051c06-1d08-41da-9c70-8769a08d3fa1" containerName="neutron-api" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.512928 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-lpdl4" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.525381 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-lpdl4"] Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.538325 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-combined-ca-bundle\") pod \"8f051c06-1d08-41da-9c70-8769a08d3fa1\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.538363 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-httpd-config\") pod \"8f051c06-1d08-41da-9c70-8769a08d3fa1\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.538419 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-config\") pod \"8f051c06-1d08-41da-9c70-8769a08d3fa1\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.538478 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-ovndb-tls-certs\") pod \"8f051c06-1d08-41da-9c70-8769a08d3fa1\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.538507 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpj8d\" (UniqueName: \"kubernetes.io/projected/8f051c06-1d08-41da-9c70-8769a08d3fa1-kube-api-access-rpj8d\") pod \"8f051c06-1d08-41da-9c70-8769a08d3fa1\" (UID: \"8f051c06-1d08-41da-9c70-8769a08d3fa1\") " Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.539023 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvn7n\" (UniqueName: \"kubernetes.io/projected/a6e71c3b-c209-4585-a6d4-e951dbf21b42-kube-api-access-dvn7n\") pod \"nova-api-db-create-fhkn7\" (UID: \"a6e71c3b-c209-4585-a6d4-e951dbf21b42\") " pod="openstack/nova-api-db-create-fhkn7" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.552949 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "8f051c06-1d08-41da-9c70-8769a08d3fa1" (UID: "8f051c06-1d08-41da-9c70-8769a08d3fa1"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.559891 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f051c06-1d08-41da-9c70-8769a08d3fa1-kube-api-access-rpj8d" (OuterVolumeSpecName: "kube-api-access-rpj8d") pod "8f051c06-1d08-41da-9c70-8769a08d3fa1" (UID: "8f051c06-1d08-41da-9c70-8769a08d3fa1"). InnerVolumeSpecName "kube-api-access-rpj8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.582831 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvn7n\" (UniqueName: \"kubernetes.io/projected/a6e71c3b-c209-4585-a6d4-e951dbf21b42-kube-api-access-dvn7n\") pod \"nova-api-db-create-fhkn7\" (UID: \"a6e71c3b-c209-4585-a6d4-e951dbf21b42\") " pod="openstack/nova-api-db-create-fhkn7" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.630148 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-qrkm9"] Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.633977 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qrkm9" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.643009 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md8ht\" (UniqueName: \"kubernetes.io/projected/5cb5981a-5bc8-4e20-b6df-77d1e5e02674-kube-api-access-md8ht\") pod \"nova-cell0-db-create-lpdl4\" (UID: \"5cb5981a-5bc8-4e20-b6df-77d1e5e02674\") " pod="openstack/nova-cell0-db-create-lpdl4" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.644331 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpj8d\" (UniqueName: \"kubernetes.io/projected/8f051c06-1d08-41da-9c70-8769a08d3fa1-kube-api-access-rpj8d\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.644363 4771 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.651743 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f051c06-1d08-41da-9c70-8769a08d3fa1" (UID: "8f051c06-1d08-41da-9c70-8769a08d3fa1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.670615 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-qrkm9"] Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.677179 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-config" (OuterVolumeSpecName: "config") pod "8f051c06-1d08-41da-9c70-8769a08d3fa1" (UID: "8f051c06-1d08-41da-9c70-8769a08d3fa1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.744933 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-fhkn7" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.748893 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4dv8\" (UniqueName: \"kubernetes.io/projected/6186a50e-d587-4972-97bf-4bb6a7a13c83-kube-api-access-t4dv8\") pod \"nova-cell1-db-create-qrkm9\" (UID: \"6186a50e-d587-4972-97bf-4bb6a7a13c83\") " pod="openstack/nova-cell1-db-create-qrkm9" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.749054 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md8ht\" (UniqueName: \"kubernetes.io/projected/5cb5981a-5bc8-4e20-b6df-77d1e5e02674-kube-api-access-md8ht\") pod \"nova-cell0-db-create-lpdl4\" (UID: \"5cb5981a-5bc8-4e20-b6df-77d1e5e02674\") " pod="openstack/nova-cell0-db-create-lpdl4" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.749254 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.749268 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.780737 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md8ht\" (UniqueName: \"kubernetes.io/projected/5cb5981a-5bc8-4e20-b6df-77d1e5e02674-kube-api-access-md8ht\") pod \"nova-cell0-db-create-lpdl4\" (UID: \"5cb5981a-5bc8-4e20-b6df-77d1e5e02674\") " pod="openstack/nova-cell0-db-create-lpdl4" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.810308 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "8f051c06-1d08-41da-9c70-8769a08d3fa1" (UID: "8f051c06-1d08-41da-9c70-8769a08d3fa1"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.851736 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4dv8\" (UniqueName: \"kubernetes.io/projected/6186a50e-d587-4972-97bf-4bb6a7a13c83-kube-api-access-t4dv8\") pod \"nova-cell1-db-create-qrkm9\" (UID: \"6186a50e-d587-4972-97bf-4bb6a7a13c83\") " pod="openstack/nova-cell1-db-create-qrkm9" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.851970 4771 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f051c06-1d08-41da-9c70-8769a08d3fa1-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.855515 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-lpdl4" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.879868 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4dv8\" (UniqueName: \"kubernetes.io/projected/6186a50e-d587-4972-97bf-4bb6a7a13c83-kube-api-access-t4dv8\") pod \"nova-cell1-db-create-qrkm9\" (UID: \"6186a50e-d587-4972-97bf-4bb6a7a13c83\") " pod="openstack/nova-cell1-db-create-qrkm9" Oct 02 10:02:21 crc kubenswrapper[4771]: I1002 10:02:21.908836 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.059326 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-scripts\") pod \"69624004-f6bf-48f2-af01-44b450260c6a\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.059485 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-config-data\") pod \"69624004-f6bf-48f2-af01-44b450260c6a\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.059548 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69624004-f6bf-48f2-af01-44b450260c6a-logs\") pod \"69624004-f6bf-48f2-af01-44b450260c6a\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.059582 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-combined-ca-bundle\") pod \"69624004-f6bf-48f2-af01-44b450260c6a\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.059610 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh4hc\" (UniqueName: \"kubernetes.io/projected/69624004-f6bf-48f2-af01-44b450260c6a-kube-api-access-jh4hc\") pod \"69624004-f6bf-48f2-af01-44b450260c6a\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.059688 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69624004-f6bf-48f2-af01-44b450260c6a-httpd-run\") pod \"69624004-f6bf-48f2-af01-44b450260c6a\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.059884 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"69624004-f6bf-48f2-af01-44b450260c6a\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.059905 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-public-tls-certs\") pod \"69624004-f6bf-48f2-af01-44b450260c6a\" (UID: \"69624004-f6bf-48f2-af01-44b450260c6a\") " Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.063475 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69624004-f6bf-48f2-af01-44b450260c6a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "69624004-f6bf-48f2-af01-44b450260c6a" (UID: "69624004-f6bf-48f2-af01-44b450260c6a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.064259 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69624004-f6bf-48f2-af01-44b450260c6a-logs" (OuterVolumeSpecName: "logs") pod "69624004-f6bf-48f2-af01-44b450260c6a" (UID: "69624004-f6bf-48f2-af01-44b450260c6a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.087429 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "69624004-f6bf-48f2-af01-44b450260c6a" (UID: "69624004-f6bf-48f2-af01-44b450260c6a"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.087643 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69624004-f6bf-48f2-af01-44b450260c6a-kube-api-access-jh4hc" (OuterVolumeSpecName: "kube-api-access-jh4hc") pod "69624004-f6bf-48f2-af01-44b450260c6a" (UID: "69624004-f6bf-48f2-af01-44b450260c6a"). InnerVolumeSpecName "kube-api-access-jh4hc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.093014 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-scripts" (OuterVolumeSpecName: "scripts") pod "69624004-f6bf-48f2-af01-44b450260c6a" (UID: "69624004-f6bf-48f2-af01-44b450260c6a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.133230 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e1a0457-04c3-4dd6-8065-a04d208525c9","Type":"ContainerStarted","Data":"9861341d50ff9c9b78d6835764f91d0fc5b0830106432d41bdeb9ba553d7a0b9"} Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.133466 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="ceilometer-central-agent" containerID="cri-o://fbe281d1e373fb76203172136537c17314552d8df3d2fa6c50de186d8b9cbd15" gracePeriod=30 Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.133726 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.134073 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="proxy-httpd" containerID="cri-o://9861341d50ff9c9b78d6835764f91d0fc5b0830106432d41bdeb9ba553d7a0b9" gracePeriod=30 Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.134119 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="sg-core" containerID="cri-o://aebc0b89a5d5628ec59f5be0429060dd4ca06f624db3f883833edf6a3e93229b" gracePeriod=30 Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.134174 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="ceilometer-notification-agent" containerID="cri-o://a3912405d27c0f84507547d3e57a9a434827c85fe0ac745c017c03df7dba93fb" gracePeriod=30 Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.168611 4771 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.168656 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.168666 4771 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69624004-f6bf-48f2-af01-44b450260c6a-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.168677 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh4hc\" (UniqueName: \"kubernetes.io/projected/69624004-f6bf-48f2-af01-44b450260c6a-kube-api-access-jh4hc\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.168688 4771 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69624004-f6bf-48f2-af01-44b450260c6a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.172851 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"69624004-f6bf-48f2-af01-44b450260c6a","Type":"ContainerDied","Data":"1cd4acf95d3a88f76237bf23f4314dd5ba1251d5f3669f036984d96178833cf8"} Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.172925 4771 scope.go:117] "RemoveContainer" containerID="8a93dc16801ca30b050caf35b39ec4c79462b62b0bc15228517efa1865e06820" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.173182 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.173435 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69624004-f6bf-48f2-af01-44b450260c6a" (UID: "69624004-f6bf-48f2-af01-44b450260c6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.174048 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qrkm9" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.175704 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.778653238 podStartE2EDuration="18.175680161s" podCreationTimestamp="2025-10-02 10:02:04 +0000 UTC" firstStartedPulling="2025-10-02 10:02:06.112737529 +0000 UTC m=+1513.760422596" lastFinishedPulling="2025-10-02 10:02:21.509764452 +0000 UTC m=+1529.157449519" observedRunningTime="2025-10-02 10:02:22.171747675 +0000 UTC m=+1529.819432742" watchObservedRunningTime="2025-10-02 10:02:22.175680161 +0000 UTC m=+1529.823365228" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.222563 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d8b77dfb-5n9ct" event={"ID":"8f051c06-1d08-41da-9c70-8769a08d3fa1","Type":"ContainerDied","Data":"a684ba0800d26c3d79eef5ee816f1fa6e219cc93dad9e8f4d9449fdcf6e392a3"} Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.222730 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d8b77dfb-5n9ct" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.231676 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "69624004-f6bf-48f2-af01-44b450260c6a" (UID: "69624004-f6bf-48f2-af01-44b450260c6a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.245009 4771 generic.go:334] "Generic (PLEG): container finished" podID="4ecc1a24-b566-4f4b-8a39-10b49533cd66" containerID="206ed65503085f9fbdac82b37985708060c35a2884d9991a7324b4d264d3228f" exitCode=0 Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.245100 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4ecc1a24-b566-4f4b-8a39-10b49533cd66","Type":"ContainerDied","Data":"206ed65503085f9fbdac82b37985708060c35a2884d9991a7324b4d264d3228f"} Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.292482 4771 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.292520 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.313315 4771 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.350613 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-config-data" (OuterVolumeSpecName: "config-data") pod "69624004-f6bf-48f2-af01-44b450260c6a" (UID: "69624004-f6bf-48f2-af01-44b450260c6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.396967 4771 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.396995 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69624004-f6bf-48f2-af01-44b450260c6a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.477440 4771 scope.go:117] "RemoveContainer" containerID="5d9dabe27276c37383ba954078c45957a44316629dc09a6b9c7f3b7d8cb06fa8" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.504211 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7d8b77dfb-5n9ct"] Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.518362 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7d8b77dfb-5n9ct"] Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.541100 4771 scope.go:117] "RemoveContainer" containerID="1a83ca34e9be89357ddfe9811bc59a269810259a0885c84ee55ea230989d8975" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.573357 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.600678 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.604254 4771 scope.go:117] "RemoveContainer" containerID="375d1f6c5d380d68ddb81d6180f95a3549ace994c00a2e19f4e9ab1801cc148a" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.643276 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:02:22 crc kubenswrapper[4771]: E1002 10:02:22.644453 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69624004-f6bf-48f2-af01-44b450260c6a" containerName="glance-log" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.644484 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="69624004-f6bf-48f2-af01-44b450260c6a" containerName="glance-log" Oct 02 10:02:22 crc kubenswrapper[4771]: E1002 10:02:22.644541 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69624004-f6bf-48f2-af01-44b450260c6a" containerName="glance-httpd" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.644551 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="69624004-f6bf-48f2-af01-44b450260c6a" containerName="glance-httpd" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.644828 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="69624004-f6bf-48f2-af01-44b450260c6a" containerName="glance-log" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.644874 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="69624004-f6bf-48f2-af01-44b450260c6a" containerName="glance-httpd" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.646759 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.649490 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.652106 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.679330 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.682981 4771 scope.go:117] "RemoveContainer" containerID="63a6291c1d4af868fb7e8bd6b0a8237a180876c54733b17a026cb296cbe0ce71" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.809785 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.809934 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.810056 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.810093 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.810402 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-scripts\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.810429 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-config-data\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.810593 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-logs\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.810625 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76r8g\" (UniqueName: \"kubernetes.io/projected/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-kube-api-access-76r8g\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.844254 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-fhkn7"] Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.914444 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-logs\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.914510 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76r8g\" (UniqueName: \"kubernetes.io/projected/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-kube-api-access-76r8g\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.914573 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.914618 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.914665 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.914686 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.914766 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-scripts\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.914793 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-config-data\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.916642 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-logs\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.917352 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.917659 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.927202 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.928732 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-scripts\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.954225 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.966323 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76r8g\" (UniqueName: \"kubernetes.io/projected/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-kube-api-access-76r8g\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:22 crc kubenswrapper[4771]: I1002 10:02:22.967974 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6c1ce83-6b12-4748-a54e-e91b4e47eb11-config-data\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.025392 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"d6c1ce83-6b12-4748-a54e-e91b4e47eb11\") " pod="openstack/glance-default-external-api-0" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.295886 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4ecc1a24-b566-4f4b-8a39-10b49533cd66","Type":"ContainerDied","Data":"509cc0aa180dbfa5b9fc470179aca0e1ccbbc18dd9bb7593d8ffb28ca2072330"} Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.295949 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="509cc0aa180dbfa5b9fc470179aca0e1ccbbc18dd9bb7593d8ffb28ca2072330" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.296496 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.299302 4771 generic.go:334] "Generic (PLEG): container finished" podID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerID="aebc0b89a5d5628ec59f5be0429060dd4ca06f624db3f883833edf6a3e93229b" exitCode=2 Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.299326 4771 generic.go:334] "Generic (PLEG): container finished" podID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerID="a3912405d27c0f84507547d3e57a9a434827c85fe0ac745c017c03df7dba93fb" exitCode=0 Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.299334 4771 generic.go:334] "Generic (PLEG): container finished" podID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerID="fbe281d1e373fb76203172136537c17314552d8df3d2fa6c50de186d8b9cbd15" exitCode=0 Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.299371 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e1a0457-04c3-4dd6-8065-a04d208525c9","Type":"ContainerDied","Data":"aebc0b89a5d5628ec59f5be0429060dd4ca06f624db3f883833edf6a3e93229b"} Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.299389 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e1a0457-04c3-4dd6-8065-a04d208525c9","Type":"ContainerDied","Data":"a3912405d27c0f84507547d3e57a9a434827c85fe0ac745c017c03df7dba93fb"} Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.299400 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e1a0457-04c3-4dd6-8065-a04d208525c9","Type":"ContainerDied","Data":"fbe281d1e373fb76203172136537c17314552d8df3d2fa6c50de186d8b9cbd15"} Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.301938 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" event={"ID":"ad488e84-ee1a-47f8-b626-fdc87919520d","Type":"ContainerStarted","Data":"83daecaf3301251049f27a0075b80ec16cf5e00ffcf3a51c9c2547cb0b86dd43"} Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.302488 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.306173 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.307753 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-fhkn7" event={"ID":"a6e71c3b-c209-4585-a6d4-e951dbf21b42","Type":"ContainerStarted","Data":"78a6be589d59d8a727e9daf7e752cde9333b9c2151402bcf99ec8f048e24adef"} Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.352363 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-lpdl4"] Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.460940 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4ecc1a24-b566-4f4b-8a39-10b49533cd66-httpd-run\") pod \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.461056 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzncw\" (UniqueName: \"kubernetes.io/projected/4ecc1a24-b566-4f4b-8a39-10b49533cd66-kube-api-access-qzncw\") pod \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.461193 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-scripts\") pod \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.461221 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-config-data\") pod \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.461244 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ecc1a24-b566-4f4b-8a39-10b49533cd66-logs\") pod \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.461336 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.461371 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-combined-ca-bundle\") pod \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.461412 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-internal-tls-certs\") pod \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\" (UID: \"4ecc1a24-b566-4f4b-8a39-10b49533cd66\") " Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.464765 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ecc1a24-b566-4f4b-8a39-10b49533cd66-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4ecc1a24-b566-4f4b-8a39-10b49533cd66" (UID: "4ecc1a24-b566-4f4b-8a39-10b49533cd66"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.473163 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ecc1a24-b566-4f4b-8a39-10b49533cd66-logs" (OuterVolumeSpecName: "logs") pod "4ecc1a24-b566-4f4b-8a39-10b49533cd66" (UID: "4ecc1a24-b566-4f4b-8a39-10b49533cd66"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.486521 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ecc1a24-b566-4f4b-8a39-10b49533cd66-kube-api-access-qzncw" (OuterVolumeSpecName: "kube-api-access-qzncw") pod "4ecc1a24-b566-4f4b-8a39-10b49533cd66" (UID: "4ecc1a24-b566-4f4b-8a39-10b49533cd66"). InnerVolumeSpecName "kube-api-access-qzncw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.498495 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "4ecc1a24-b566-4f4b-8a39-10b49533cd66" (UID: "4ecc1a24-b566-4f4b-8a39-10b49533cd66"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.508997 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-scripts" (OuterVolumeSpecName: "scripts") pod "4ecc1a24-b566-4f4b-8a39-10b49533cd66" (UID: "4ecc1a24-b566-4f4b-8a39-10b49533cd66"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.566462 4771 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4ecc1a24-b566-4f4b-8a39-10b49533cd66-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.566937 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzncw\" (UniqueName: \"kubernetes.io/projected/4ecc1a24-b566-4f4b-8a39-10b49533cd66-kube-api-access-qzncw\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.566948 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.566958 4771 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ecc1a24-b566-4f4b-8a39-10b49533cd66-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.567084 4771 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.693853 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ecc1a24-b566-4f4b-8a39-10b49533cd66" (UID: "4ecc1a24-b566-4f4b-8a39-10b49533cd66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.765354 4771 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.777551 4771 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.777590 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.836251 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69624004-f6bf-48f2-af01-44b450260c6a" path="/var/lib/kubelet/pods/69624004-f6bf-48f2-af01-44b450260c6a/volumes" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.838984 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f051c06-1d08-41da-9c70-8769a08d3fa1" path="/var/lib/kubelet/pods/8f051c06-1d08-41da-9c70-8769a08d3fa1/volumes" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.849585 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-config-data" (OuterVolumeSpecName: "config-data") pod "4ecc1a24-b566-4f4b-8a39-10b49533cd66" (UID: "4ecc1a24-b566-4f4b-8a39-10b49533cd66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.854893 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4ecc1a24-b566-4f4b-8a39-10b49533cd66" (UID: "4ecc1a24-b566-4f4b-8a39-10b49533cd66"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.874735 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-qrkm9"] Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.888036 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:23 crc kubenswrapper[4771]: I1002 10:02:23.888082 4771 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ecc1a24-b566-4f4b-8a39-10b49533cd66-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:23 crc kubenswrapper[4771]: W1002 10:02:23.897637 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6186a50e_d587_4972_97bf_4bb6a7a13c83.slice/crio-0efaa7425016f284a9ebeff1534ad15bd8ae087c0deeed9f7d51357911cde8be WatchSource:0}: Error finding container 0efaa7425016f284a9ebeff1534ad15bd8ae087c0deeed9f7d51357911cde8be: Status 404 returned error can't find the container with id 0efaa7425016f284a9ebeff1534ad15bd8ae087c0deeed9f7d51357911cde8be Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.178417 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:02:24 crc kubenswrapper[4771]: W1002 10:02:24.180687 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6c1ce83_6b12_4748_a54e_e91b4e47eb11.slice/crio-ebd7ab2a1557471aa9a424d909abfe6f23194edd76e7941ab78a4df3f7ae83cb WatchSource:0}: Error finding container ebd7ab2a1557471aa9a424d909abfe6f23194edd76e7941ab78a4df3f7ae83cb: Status 404 returned error can't find the container with id ebd7ab2a1557471aa9a424d909abfe6f23194edd76e7941ab78a4df3f7ae83cb Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.370108 4771 generic.go:334] "Generic (PLEG): container finished" podID="a6e71c3b-c209-4585-a6d4-e951dbf21b42" containerID="453ad515e7c0ffc2b998189de46236c802c1791795cd8dd14d6bf8f2b7e3a0ee" exitCode=0 Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.370551 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-fhkn7" event={"ID":"a6e71c3b-c209-4585-a6d4-e951dbf21b42","Type":"ContainerDied","Data":"453ad515e7c0ffc2b998189de46236c802c1791795cd8dd14d6bf8f2b7e3a0ee"} Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.382398 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d6c1ce83-6b12-4748-a54e-e91b4e47eb11","Type":"ContainerStarted","Data":"ebd7ab2a1557471aa9a424d909abfe6f23194edd76e7941ab78a4df3f7ae83cb"} Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.419784 4771 generic.go:334] "Generic (PLEG): container finished" podID="5cb5981a-5bc8-4e20-b6df-77d1e5e02674" containerID="2830b607f8ef44d30d755763b6bd0ebe201f45c32fd97f40a505528eea8a69e2" exitCode=0 Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.419880 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-lpdl4" event={"ID":"5cb5981a-5bc8-4e20-b6df-77d1e5e02674","Type":"ContainerDied","Data":"2830b607f8ef44d30d755763b6bd0ebe201f45c32fd97f40a505528eea8a69e2"} Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.419951 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-lpdl4" event={"ID":"5cb5981a-5bc8-4e20-b6df-77d1e5e02674","Type":"ContainerStarted","Data":"6278a917b0976a3c7070d7f03b86b573b03b5c32611f53e7b45e5b41c06bc0bf"} Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.428442 4771 generic.go:334] "Generic (PLEG): container finished" podID="ad488e84-ee1a-47f8-b626-fdc87919520d" containerID="83daecaf3301251049f27a0075b80ec16cf5e00ffcf3a51c9c2547cb0b86dd43" exitCode=1 Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.429251 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" event={"ID":"ad488e84-ee1a-47f8-b626-fdc87919520d","Type":"ContainerDied","Data":"83daecaf3301251049f27a0075b80ec16cf5e00ffcf3a51c9c2547cb0b86dd43"} Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.429363 4771 scope.go:117] "RemoveContainer" containerID="63a6291c1d4af868fb7e8bd6b0a8237a180876c54733b17a026cb296cbe0ce71" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.429586 4771 scope.go:117] "RemoveContainer" containerID="83daecaf3301251049f27a0075b80ec16cf5e00ffcf3a51c9c2547cb0b86dd43" Oct 02 10:02:24 crc kubenswrapper[4771]: E1002 10:02:24.430041 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 20s restarting failed container=heat-cfnapi pod=heat-cfnapi-5d49fd5cb-kh9dz_openstack(ad488e84-ee1a-47f8-b626-fdc87919520d)\"" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.446498 4771 generic.go:334] "Generic (PLEG): container finished" podID="6186a50e-d587-4972-97bf-4bb6a7a13c83" containerID="ddf68268fe2909555075263c92dc36a09e66dd98d090846f1f252ee4442c17d1" exitCode=0 Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.446613 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.448177 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qrkm9" event={"ID":"6186a50e-d587-4972-97bf-4bb6a7a13c83","Type":"ContainerDied","Data":"ddf68268fe2909555075263c92dc36a09e66dd98d090846f1f252ee4442c17d1"} Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.448228 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qrkm9" event={"ID":"6186a50e-d587-4972-97bf-4bb6a7a13c83","Type":"ContainerStarted","Data":"0efaa7425016f284a9ebeff1534ad15bd8ae087c0deeed9f7d51357911cde8be"} Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.603374 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.612831 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.642772 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.672326 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:02:24 crc kubenswrapper[4771]: E1002 10:02:24.672976 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ecc1a24-b566-4f4b-8a39-10b49533cd66" containerName="glance-httpd" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.672994 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ecc1a24-b566-4f4b-8a39-10b49533cd66" containerName="glance-httpd" Oct 02 10:02:24 crc kubenswrapper[4771]: E1002 10:02:24.673073 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ecc1a24-b566-4f4b-8a39-10b49533cd66" containerName="glance-log" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.673082 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ecc1a24-b566-4f4b-8a39-10b49533cd66" containerName="glance-log" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.673403 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ecc1a24-b566-4f4b-8a39-10b49533cd66" containerName="glance-httpd" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.673442 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ecc1a24-b566-4f4b-8a39-10b49533cd66" containerName="glance-log" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.674864 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.680436 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.680804 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.683115 4771 scope.go:117] "RemoveContainer" containerID="f3bd35ae1d18f4913d4344dba97cb1172d9600dd4b838f23985dd92ef30a1368" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.690968 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.836525 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/606540a4-01ff-4915-89b2-f2a72c2fa922-config-data\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.837070 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/606540a4-01ff-4915-89b2-f2a72c2fa922-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.837219 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/606540a4-01ff-4915-89b2-f2a72c2fa922-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.837279 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xcx4\" (UniqueName: \"kubernetes.io/projected/606540a4-01ff-4915-89b2-f2a72c2fa922-kube-api-access-2xcx4\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.837377 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/606540a4-01ff-4915-89b2-f2a72c2fa922-logs\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.837415 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/606540a4-01ff-4915-89b2-f2a72c2fa922-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.837477 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.837563 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/606540a4-01ff-4915-89b2-f2a72c2fa922-scripts\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.940596 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/606540a4-01ff-4915-89b2-f2a72c2fa922-config-data\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.940653 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/606540a4-01ff-4915-89b2-f2a72c2fa922-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.940748 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/606540a4-01ff-4915-89b2-f2a72c2fa922-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.940810 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xcx4\" (UniqueName: \"kubernetes.io/projected/606540a4-01ff-4915-89b2-f2a72c2fa922-kube-api-access-2xcx4\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.940889 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/606540a4-01ff-4915-89b2-f2a72c2fa922-logs\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.940932 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/606540a4-01ff-4915-89b2-f2a72c2fa922-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.940994 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.941068 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/606540a4-01ff-4915-89b2-f2a72c2fa922-scripts\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.945579 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.946514 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/606540a4-01ff-4915-89b2-f2a72c2fa922-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.953416 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/606540a4-01ff-4915-89b2-f2a72c2fa922-config-data\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.956102 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/606540a4-01ff-4915-89b2-f2a72c2fa922-scripts\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.960005 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/606540a4-01ff-4915-89b2-f2a72c2fa922-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.960561 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/606540a4-01ff-4915-89b2-f2a72c2fa922-logs\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.967446 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/606540a4-01ff-4915-89b2-f2a72c2fa922-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:24 crc kubenswrapper[4771]: I1002 10:02:24.969295 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xcx4\" (UniqueName: \"kubernetes.io/projected/606540a4-01ff-4915-89b2-f2a72c2fa922-kube-api-access-2xcx4\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:25 crc kubenswrapper[4771]: I1002 10:02:25.028705 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"606540a4-01ff-4915-89b2-f2a72c2fa922\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:02:25 crc kubenswrapper[4771]: I1002 10:02:25.313389 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:02:25 crc kubenswrapper[4771]: I1002 10:02:25.463568 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8ff6c7648-wtgnm" event={"ID":"983d8aca-7ee4-4105-9678-63d608645975","Type":"ContainerStarted","Data":"bcbeb31e7f398e9be0ad8b2e71485ba7b2c694c8b9de4cb550a91527ea849700"} Oct 02 10:02:25 crc kubenswrapper[4771]: I1002 10:02:25.463797 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:02:25 crc kubenswrapper[4771]: I1002 10:02:25.472591 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d6c1ce83-6b12-4748-a54e-e91b4e47eb11","Type":"ContainerStarted","Data":"59c19027079bd844dabf41345bd65ddbfd1a37e4b575224d1dace123ed155641"} Oct 02 10:02:25 crc kubenswrapper[4771]: I1002 10:02:25.476193 4771 scope.go:117] "RemoveContainer" containerID="83daecaf3301251049f27a0075b80ec16cf5e00ffcf3a51c9c2547cb0b86dd43" Oct 02 10:02:25 crc kubenswrapper[4771]: E1002 10:02:25.476711 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 20s restarting failed container=heat-cfnapi pod=heat-cfnapi-5d49fd5cb-kh9dz_openstack(ad488e84-ee1a-47f8-b626-fdc87919520d)\"" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" Oct 02 10:02:25 crc kubenswrapper[4771]: I1002 10:02:25.715891 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ecc1a24-b566-4f4b-8a39-10b49533cd66" path="/var/lib/kubelet/pods/4ecc1a24-b566-4f4b-8a39-10b49533cd66/volumes" Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.513423 4771 generic.go:334] "Generic (PLEG): container finished" podID="983d8aca-7ee4-4105-9678-63d608645975" containerID="bcbeb31e7f398e9be0ad8b2e71485ba7b2c694c8b9de4cb550a91527ea849700" exitCode=1 Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.513839 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8ff6c7648-wtgnm" event={"ID":"983d8aca-7ee4-4105-9678-63d608645975","Type":"ContainerDied","Data":"bcbeb31e7f398e9be0ad8b2e71485ba7b2c694c8b9de4cb550a91527ea849700"} Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.513887 4771 scope.go:117] "RemoveContainer" containerID="f3bd35ae1d18f4913d4344dba97cb1172d9600dd4b838f23985dd92ef30a1368" Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.515317 4771 scope.go:117] "RemoveContainer" containerID="bcbeb31e7f398e9be0ad8b2e71485ba7b2c694c8b9de4cb550a91527ea849700" Oct 02 10:02:26 crc kubenswrapper[4771]: E1002 10:02:26.515984 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 20s restarting failed container=heat-api pod=heat-api-8ff6c7648-wtgnm_openstack(983d8aca-7ee4-4105-9678-63d608645975)\"" pod="openstack/heat-api-8ff6c7648-wtgnm" podUID="983d8aca-7ee4-4105-9678-63d608645975" Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.551065 4771 scope.go:117] "RemoveContainer" containerID="83daecaf3301251049f27a0075b80ec16cf5e00ffcf3a51c9c2547cb0b86dd43" Oct 02 10:02:26 crc kubenswrapper[4771]: E1002 10:02:26.551387 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 20s restarting failed container=heat-cfnapi pod=heat-cfnapi-5d49fd5cb-kh9dz_openstack(ad488e84-ee1a-47f8-b626-fdc87919520d)\"" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.551736 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d6c1ce83-6b12-4748-a54e-e91b4e47eb11","Type":"ContainerStarted","Data":"6ae6905daa30ff16f8300076d9a3d4aa694a8b14ce4470628c2253891a8e61b3"} Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.554556 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:02:26 crc kubenswrapper[4771]: W1002 10:02:26.559046 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod606540a4_01ff_4915_89b2_f2a72c2fa922.slice/crio-aa2fe9750c3ab8913a3523fd7e45c867d4795bfc785728920cab95b31a994207 WatchSource:0}: Error finding container aa2fe9750c3ab8913a3523fd7e45c867d4795bfc785728920cab95b31a994207: Status 404 returned error can't find the container with id aa2fe9750c3ab8913a3523fd7e45c867d4795bfc785728920cab95b31a994207 Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.599816 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.5997901070000005 podStartE2EDuration="4.599790107s" podCreationTimestamp="2025-10-02 10:02:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:02:26.592883401 +0000 UTC m=+1534.240568488" watchObservedRunningTime="2025-10-02 10:02:26.599790107 +0000 UTC m=+1534.247475174" Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.764749 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qrkm9" Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.765825 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-lpdl4" Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.778472 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-fhkn7" Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.964454 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md8ht\" (UniqueName: \"kubernetes.io/projected/5cb5981a-5bc8-4e20-b6df-77d1e5e02674-kube-api-access-md8ht\") pod \"5cb5981a-5bc8-4e20-b6df-77d1e5e02674\" (UID: \"5cb5981a-5bc8-4e20-b6df-77d1e5e02674\") " Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.964971 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4dv8\" (UniqueName: \"kubernetes.io/projected/6186a50e-d587-4972-97bf-4bb6a7a13c83-kube-api-access-t4dv8\") pod \"6186a50e-d587-4972-97bf-4bb6a7a13c83\" (UID: \"6186a50e-d587-4972-97bf-4bb6a7a13c83\") " Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.965431 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvn7n\" (UniqueName: \"kubernetes.io/projected/a6e71c3b-c209-4585-a6d4-e951dbf21b42-kube-api-access-dvn7n\") pod \"a6e71c3b-c209-4585-a6d4-e951dbf21b42\" (UID: \"a6e71c3b-c209-4585-a6d4-e951dbf21b42\") " Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.972728 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6186a50e-d587-4972-97bf-4bb6a7a13c83-kube-api-access-t4dv8" (OuterVolumeSpecName: "kube-api-access-t4dv8") pod "6186a50e-d587-4972-97bf-4bb6a7a13c83" (UID: "6186a50e-d587-4972-97bf-4bb6a7a13c83"). InnerVolumeSpecName "kube-api-access-t4dv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.973156 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6e71c3b-c209-4585-a6d4-e951dbf21b42-kube-api-access-dvn7n" (OuterVolumeSpecName: "kube-api-access-dvn7n") pod "a6e71c3b-c209-4585-a6d4-e951dbf21b42" (UID: "a6e71c3b-c209-4585-a6d4-e951dbf21b42"). InnerVolumeSpecName "kube-api-access-dvn7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:26 crc kubenswrapper[4771]: I1002 10:02:26.976252 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cb5981a-5bc8-4e20-b6df-77d1e5e02674-kube-api-access-md8ht" (OuterVolumeSpecName: "kube-api-access-md8ht") pod "5cb5981a-5bc8-4e20-b6df-77d1e5e02674" (UID: "5cb5981a-5bc8-4e20-b6df-77d1e5e02674"). InnerVolumeSpecName "kube-api-access-md8ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.068937 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4dv8\" (UniqueName: \"kubernetes.io/projected/6186a50e-d587-4972-97bf-4bb6a7a13c83-kube-api-access-t4dv8\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.068978 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvn7n\" (UniqueName: \"kubernetes.io/projected/a6e71c3b-c209-4585-a6d4-e951dbf21b42-kube-api-access-dvn7n\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.068987 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md8ht\" (UniqueName: \"kubernetes.io/projected/5cb5981a-5bc8-4e20-b6df-77d1e5e02674-kube-api-access-md8ht\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.564674 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qrkm9" event={"ID":"6186a50e-d587-4972-97bf-4bb6a7a13c83","Type":"ContainerDied","Data":"0efaa7425016f284a9ebeff1534ad15bd8ae087c0deeed9f7d51357911cde8be"} Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.565084 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0efaa7425016f284a9ebeff1534ad15bd8ae087c0deeed9f7d51357911cde8be" Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.564719 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qrkm9" Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.566971 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"606540a4-01ff-4915-89b2-f2a72c2fa922","Type":"ContainerStarted","Data":"60d7777b8bf47199e32f25fcea8a4e1d907d3d224165dac7c955c21f28b936ba"} Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.567019 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"606540a4-01ff-4915-89b2-f2a72c2fa922","Type":"ContainerStarted","Data":"aa2fe9750c3ab8913a3523fd7e45c867d4795bfc785728920cab95b31a994207"} Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.572820 4771 scope.go:117] "RemoveContainer" containerID="bcbeb31e7f398e9be0ad8b2e71485ba7b2c694c8b9de4cb550a91527ea849700" Oct 02 10:02:27 crc kubenswrapper[4771]: E1002 10:02:27.573064 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 20s restarting failed container=heat-api pod=heat-api-8ff6c7648-wtgnm_openstack(983d8aca-7ee4-4105-9678-63d608645975)\"" pod="openstack/heat-api-8ff6c7648-wtgnm" podUID="983d8aca-7ee4-4105-9678-63d608645975" Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.576006 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-fhkn7" Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.576010 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-fhkn7" event={"ID":"a6e71c3b-c209-4585-a6d4-e951dbf21b42","Type":"ContainerDied","Data":"78a6be589d59d8a727e9daf7e752cde9333b9c2151402bcf99ec8f048e24adef"} Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.576063 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78a6be589d59d8a727e9daf7e752cde9333b9c2151402bcf99ec8f048e24adef" Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.583798 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-lpdl4" Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.584035 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-lpdl4" event={"ID":"5cb5981a-5bc8-4e20-b6df-77d1e5e02674","Type":"ContainerDied","Data":"6278a917b0976a3c7070d7f03b86b573b03b5c32611f53e7b45e5b41c06bc0bf"} Oct 02 10:02:27 crc kubenswrapper[4771]: I1002 10:02:27.585049 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6278a917b0976a3c7070d7f03b86b573b03b5c32611f53e7b45e5b41c06bc0bf" Oct 02 10:02:28 crc kubenswrapper[4771]: E1002 10:02:28.191249 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6e71c3b_c209_4585_a6d4_e951dbf21b42.slice/crio-78a6be589d59d8a727e9daf7e752cde9333b9c2151402bcf99ec8f048e24adef\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6e71c3b_c209_4585_a6d4_e951dbf21b42.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cb5981a_5bc8_4e20_b6df_77d1e5e02674.slice/crio-6278a917b0976a3c7070d7f03b86b573b03b5c32611f53e7b45e5b41c06bc0bf\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice/crio-8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice\": RecentStats: unable to find data in memory cache]" Oct 02 10:02:28 crc kubenswrapper[4771]: I1002 10:02:28.598749 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"606540a4-01ff-4915-89b2-f2a72c2fa922","Type":"ContainerStarted","Data":"4cbdddac658547e78c6e236218d0b3a9cf5337dd4b1ab084566486d1341739f0"} Oct 02 10:02:28 crc kubenswrapper[4771]: I1002 10:02:28.633880 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.633858944 podStartE2EDuration="4.633858944s" podCreationTimestamp="2025-10-02 10:02:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:02:28.627621086 +0000 UTC m=+1536.275306153" watchObservedRunningTime="2025-10-02 10:02:28.633858944 +0000 UTC m=+1536.281544001" Oct 02 10:02:29 crc kubenswrapper[4771]: I1002 10:02:29.408643 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:02:29 crc kubenswrapper[4771]: I1002 10:02:29.491941 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5d49fd5cb-kh9dz"] Oct 02 10:02:29 crc kubenswrapper[4771]: I1002 10:02:29.574670 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:02:29 crc kubenswrapper[4771]: I1002 10:02:29.575593 4771 scope.go:117] "RemoveContainer" containerID="bcbeb31e7f398e9be0ad8b2e71485ba7b2c694c8b9de4cb550a91527ea849700" Oct 02 10:02:29 crc kubenswrapper[4771]: E1002 10:02:29.576568 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 20s restarting failed container=heat-api pod=heat-api-8ff6c7648-wtgnm_openstack(983d8aca-7ee4-4105-9678-63d608645975)\"" pod="openstack/heat-api-8ff6c7648-wtgnm" podUID="983d8aca-7ee4-4105-9678-63d608645975" Oct 02 10:02:29 crc kubenswrapper[4771]: I1002 10:02:29.593853 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:02:29 crc kubenswrapper[4771]: I1002 10:02:29.709151 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-8ff6c7648-wtgnm"] Oct 02 10:02:30 crc kubenswrapper[4771]: E1002 10:02:30.341543 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 02 10:02:30 crc kubenswrapper[4771]: E1002 10:02:30.343859 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 02 10:02:30 crc kubenswrapper[4771]: E1002 10:02:30.347572 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 02 10:02:30 crc kubenswrapper[4771]: E1002 10:02:30.347663 4771 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-649769f8cf-k8cph" podUID="d7b40340-a806-490f-9842-a074ffba2e93" containerName="heat-engine" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.436049 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.443082 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.498600 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlzgp\" (UniqueName: \"kubernetes.io/projected/ad488e84-ee1a-47f8-b626-fdc87919520d-kube-api-access-vlzgp\") pod \"ad488e84-ee1a-47f8-b626-fdc87919520d\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.498706 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-config-data\") pod \"ad488e84-ee1a-47f8-b626-fdc87919520d\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.498949 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-config-data\") pod \"983d8aca-7ee4-4105-9678-63d608645975\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.499009 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-config-data-custom\") pod \"983d8aca-7ee4-4105-9678-63d608645975\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.499054 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-config-data-custom\") pod \"ad488e84-ee1a-47f8-b626-fdc87919520d\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.499132 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-combined-ca-bundle\") pod \"983d8aca-7ee4-4105-9678-63d608645975\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.499192 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s47k6\" (UniqueName: \"kubernetes.io/projected/983d8aca-7ee4-4105-9678-63d608645975-kube-api-access-s47k6\") pod \"983d8aca-7ee4-4105-9678-63d608645975\" (UID: \"983d8aca-7ee4-4105-9678-63d608645975\") " Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.499256 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-combined-ca-bundle\") pod \"ad488e84-ee1a-47f8-b626-fdc87919520d\" (UID: \"ad488e84-ee1a-47f8-b626-fdc87919520d\") " Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.509070 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad488e84-ee1a-47f8-b626-fdc87919520d-kube-api-access-vlzgp" (OuterVolumeSpecName: "kube-api-access-vlzgp") pod "ad488e84-ee1a-47f8-b626-fdc87919520d" (UID: "ad488e84-ee1a-47f8-b626-fdc87919520d"). InnerVolumeSpecName "kube-api-access-vlzgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.510109 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ad488e84-ee1a-47f8-b626-fdc87919520d" (UID: "ad488e84-ee1a-47f8-b626-fdc87919520d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.511204 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "983d8aca-7ee4-4105-9678-63d608645975" (UID: "983d8aca-7ee4-4105-9678-63d608645975"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.511902 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/983d8aca-7ee4-4105-9678-63d608645975-kube-api-access-s47k6" (OuterVolumeSpecName: "kube-api-access-s47k6") pod "983d8aca-7ee4-4105-9678-63d608645975" (UID: "983d8aca-7ee4-4105-9678-63d608645975"). InnerVolumeSpecName "kube-api-access-s47k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.575773 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad488e84-ee1a-47f8-b626-fdc87919520d" (UID: "ad488e84-ee1a-47f8-b626-fdc87919520d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.586305 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "983d8aca-7ee4-4105-9678-63d608645975" (UID: "983d8aca-7ee4-4105-9678-63d608645975"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.605171 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlzgp\" (UniqueName: \"kubernetes.io/projected/ad488e84-ee1a-47f8-b626-fdc87919520d-kube-api-access-vlzgp\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.605241 4771 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.605260 4771 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.605273 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.605285 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s47k6\" (UniqueName: \"kubernetes.io/projected/983d8aca-7ee4-4105-9678-63d608645975-kube-api-access-s47k6\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.605323 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.605857 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-config-data" (OuterVolumeSpecName: "config-data") pod "983d8aca-7ee4-4105-9678-63d608645975" (UID: "983d8aca-7ee4-4105-9678-63d608645975"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.622346 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-config-data" (OuterVolumeSpecName: "config-data") pod "ad488e84-ee1a-47f8-b626-fdc87919520d" (UID: "ad488e84-ee1a-47f8-b626-fdc87919520d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.629804 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8ff6c7648-wtgnm" event={"ID":"983d8aca-7ee4-4105-9678-63d608645975","Type":"ContainerDied","Data":"f37189e1dd8f9f61d420c868f671916ae3625cb4ac79e5bf9b52138076066217"} Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.629871 4771 scope.go:117] "RemoveContainer" containerID="bcbeb31e7f398e9be0ad8b2e71485ba7b2c694c8b9de4cb550a91527ea849700" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.630001 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8ff6c7648-wtgnm" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.642227 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" event={"ID":"ad488e84-ee1a-47f8-b626-fdc87919520d","Type":"ContainerDied","Data":"36d8e029971d773d213042f48869ced689910f3bfceb01d6a62d43f0af2b7b6e"} Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.642323 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5d49fd5cb-kh9dz" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.705229 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-8ff6c7648-wtgnm"] Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.708647 4771 scope.go:117] "RemoveContainer" containerID="83daecaf3301251049f27a0075b80ec16cf5e00ffcf3a51c9c2547cb0b86dd43" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.709188 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad488e84-ee1a-47f8-b626-fdc87919520d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.709217 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983d8aca-7ee4-4105-9678-63d608645975-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.720076 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-8ff6c7648-wtgnm"] Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.730667 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5d49fd5cb-kh9dz"] Oct 02 10:02:30 crc kubenswrapper[4771]: I1002 10:02:30.750462 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-5d49fd5cb-kh9dz"] Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.529544 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0ebb-account-create-sv52h"] Oct 02 10:02:31 crc kubenswrapper[4771]: E1002 10:02:31.530265 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="983d8aca-7ee4-4105-9678-63d608645975" containerName="heat-api" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530289 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="983d8aca-7ee4-4105-9678-63d608645975" containerName="heat-api" Oct 02 10:02:31 crc kubenswrapper[4771]: E1002 10:02:31.530315 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" containerName="heat-cfnapi" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530323 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" containerName="heat-cfnapi" Oct 02 10:02:31 crc kubenswrapper[4771]: E1002 10:02:31.530335 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="983d8aca-7ee4-4105-9678-63d608645975" containerName="heat-api" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530342 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="983d8aca-7ee4-4105-9678-63d608645975" containerName="heat-api" Oct 02 10:02:31 crc kubenswrapper[4771]: E1002 10:02:31.530358 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" containerName="heat-cfnapi" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530365 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" containerName="heat-cfnapi" Oct 02 10:02:31 crc kubenswrapper[4771]: E1002 10:02:31.530389 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" containerName="heat-cfnapi" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530398 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" containerName="heat-cfnapi" Oct 02 10:02:31 crc kubenswrapper[4771]: E1002 10:02:31.530421 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="983d8aca-7ee4-4105-9678-63d608645975" containerName="heat-api" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530427 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="983d8aca-7ee4-4105-9678-63d608645975" containerName="heat-api" Oct 02 10:02:31 crc kubenswrapper[4771]: E1002 10:02:31.530443 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb5981a-5bc8-4e20-b6df-77d1e5e02674" containerName="mariadb-database-create" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530451 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb5981a-5bc8-4e20-b6df-77d1e5e02674" containerName="mariadb-database-create" Oct 02 10:02:31 crc kubenswrapper[4771]: E1002 10:02:31.530463 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6e71c3b-c209-4585-a6d4-e951dbf21b42" containerName="mariadb-database-create" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530469 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6e71c3b-c209-4585-a6d4-e951dbf21b42" containerName="mariadb-database-create" Oct 02 10:02:31 crc kubenswrapper[4771]: E1002 10:02:31.530478 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6186a50e-d587-4972-97bf-4bb6a7a13c83" containerName="mariadb-database-create" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530484 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6186a50e-d587-4972-97bf-4bb6a7a13c83" containerName="mariadb-database-create" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530718 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="983d8aca-7ee4-4105-9678-63d608645975" containerName="heat-api" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530730 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb5981a-5bc8-4e20-b6df-77d1e5e02674" containerName="mariadb-database-create" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530740 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" containerName="heat-cfnapi" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530755 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="983d8aca-7ee4-4105-9678-63d608645975" containerName="heat-api" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530766 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" containerName="heat-cfnapi" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530778 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="983d8aca-7ee4-4105-9678-63d608645975" containerName="heat-api" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530795 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="6186a50e-d587-4972-97bf-4bb6a7a13c83" containerName="mariadb-database-create" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.530813 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6e71c3b-c209-4585-a6d4-e951dbf21b42" containerName="mariadb-database-create" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.531791 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0ebb-account-create-sv52h" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.534684 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.556635 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0ebb-account-create-sv52h"] Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.632914 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf7hm\" (UniqueName: \"kubernetes.io/projected/4bbca829-8863-481a-8ff3-661dc1bc9726-kube-api-access-vf7hm\") pod \"nova-api-0ebb-account-create-sv52h\" (UID: \"4bbca829-8863-481a-8ff3-661dc1bc9726\") " pod="openstack/nova-api-0ebb-account-create-sv52h" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.735023 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="983d8aca-7ee4-4105-9678-63d608645975" path="/var/lib/kubelet/pods/983d8aca-7ee4-4105-9678-63d608645975/volumes" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.735902 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" path="/var/lib/kubelet/pods/ad488e84-ee1a-47f8-b626-fdc87919520d/volumes" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.735963 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf7hm\" (UniqueName: \"kubernetes.io/projected/4bbca829-8863-481a-8ff3-661dc1bc9726-kube-api-access-vf7hm\") pod \"nova-api-0ebb-account-create-sv52h\" (UID: \"4bbca829-8863-481a-8ff3-661dc1bc9726\") " pod="openstack/nova-api-0ebb-account-create-sv52h" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.749922 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-fc64-account-create-c4fs5"] Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.750983 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad488e84-ee1a-47f8-b626-fdc87919520d" containerName="heat-cfnapi" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.757019 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fc64-account-create-c4fs5" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.765721 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-fc64-account-create-c4fs5"] Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.778621 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.780333 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf7hm\" (UniqueName: \"kubernetes.io/projected/4bbca829-8863-481a-8ff3-661dc1bc9726-kube-api-access-vf7hm\") pod \"nova-api-0ebb-account-create-sv52h\" (UID: \"4bbca829-8863-481a-8ff3-661dc1bc9726\") " pod="openstack/nova-api-0ebb-account-create-sv52h" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.837932 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf2nt\" (UniqueName: \"kubernetes.io/projected/2cb58c29-0f39-4260-85ce-9a93e3b90d99-kube-api-access-sf2nt\") pod \"nova-cell0-fc64-account-create-c4fs5\" (UID: \"2cb58c29-0f39-4260-85ce-9a93e3b90d99\") " pod="openstack/nova-cell0-fc64-account-create-c4fs5" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.851385 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0ebb-account-create-sv52h" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.921662 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-053a-account-create-jt258"] Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.927513 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-053a-account-create-jt258" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.932390 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.939912 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf2nt\" (UniqueName: \"kubernetes.io/projected/2cb58c29-0f39-4260-85ce-9a93e3b90d99-kube-api-access-sf2nt\") pod \"nova-cell0-fc64-account-create-c4fs5\" (UID: \"2cb58c29-0f39-4260-85ce-9a93e3b90d99\") " pod="openstack/nova-cell0-fc64-account-create-c4fs5" Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.963512 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-053a-account-create-jt258"] Oct 02 10:02:31 crc kubenswrapper[4771]: I1002 10:02:31.980102 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf2nt\" (UniqueName: \"kubernetes.io/projected/2cb58c29-0f39-4260-85ce-9a93e3b90d99-kube-api-access-sf2nt\") pod \"nova-cell0-fc64-account-create-c4fs5\" (UID: \"2cb58c29-0f39-4260-85ce-9a93e3b90d99\") " pod="openstack/nova-cell0-fc64-account-create-c4fs5" Oct 02 10:02:32 crc kubenswrapper[4771]: I1002 10:02:32.041746 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwzkr\" (UniqueName: \"kubernetes.io/projected/8d3458e9-340f-43bf-a070-7e3ec1250176-kube-api-access-wwzkr\") pod \"nova-cell1-053a-account-create-jt258\" (UID: \"8d3458e9-340f-43bf-a070-7e3ec1250176\") " pod="openstack/nova-cell1-053a-account-create-jt258" Oct 02 10:02:32 crc kubenswrapper[4771]: I1002 10:02:32.144852 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwzkr\" (UniqueName: \"kubernetes.io/projected/8d3458e9-340f-43bf-a070-7e3ec1250176-kube-api-access-wwzkr\") pod \"nova-cell1-053a-account-create-jt258\" (UID: \"8d3458e9-340f-43bf-a070-7e3ec1250176\") " pod="openstack/nova-cell1-053a-account-create-jt258" Oct 02 10:02:32 crc kubenswrapper[4771]: I1002 10:02:32.157492 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fc64-account-create-c4fs5" Oct 02 10:02:32 crc kubenswrapper[4771]: I1002 10:02:32.167826 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwzkr\" (UniqueName: \"kubernetes.io/projected/8d3458e9-340f-43bf-a070-7e3ec1250176-kube-api-access-wwzkr\") pod \"nova-cell1-053a-account-create-jt258\" (UID: \"8d3458e9-340f-43bf-a070-7e3ec1250176\") " pod="openstack/nova-cell1-053a-account-create-jt258" Oct 02 10:02:32 crc kubenswrapper[4771]: E1002 10:02:32.417519 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice/crio-8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f\": RecentStats: unable to find data in memory cache]" Oct 02 10:02:32 crc kubenswrapper[4771]: I1002 10:02:32.434469 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-053a-account-create-jt258" Oct 02 10:02:32 crc kubenswrapper[4771]: I1002 10:02:32.491832 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0ebb-account-create-sv52h"] Oct 02 10:02:32 crc kubenswrapper[4771]: I1002 10:02:32.696067 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0ebb-account-create-sv52h" event={"ID":"4bbca829-8863-481a-8ff3-661dc1bc9726","Type":"ContainerStarted","Data":"c4a1f386287aba9d2944dc5fe5fb03cbaf7eb4b2f211b430d31441904ae05ce1"} Oct 02 10:02:32 crc kubenswrapper[4771]: I1002 10:02:32.739354 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-fc64-account-create-c4fs5"] Oct 02 10:02:32 crc kubenswrapper[4771]: W1002 10:02:32.751406 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2cb58c29_0f39_4260_85ce_9a93e3b90d99.slice/crio-74afd284f0ffefd3ec35d1d4bd4b4bb6b6505ed078366c53f1006b3805c67120 WatchSource:0}: Error finding container 74afd284f0ffefd3ec35d1d4bd4b4bb6b6505ed078366c53f1006b3805c67120: Status 404 returned error can't find the container with id 74afd284f0ffefd3ec35d1d4bd4b4bb6b6505ed078366c53f1006b3805c67120 Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.140512 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-053a-account-create-jt258"] Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.306404 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.306463 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.387418 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.407520 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.726196 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.728722 4771 generic.go:334] "Generic (PLEG): container finished" podID="d7b40340-a806-490f-9842-a074ffba2e93" containerID="d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98" exitCode=0 Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.728796 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-649769f8cf-k8cph" event={"ID":"d7b40340-a806-490f-9842-a074ffba2e93","Type":"ContainerDied","Data":"d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98"} Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.728830 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-649769f8cf-k8cph" event={"ID":"d7b40340-a806-490f-9842-a074ffba2e93","Type":"ContainerDied","Data":"04990be5940d31e6b9fe743a99f836720a02dcd7bd81ce0a745c3a5c448c5ef9"} Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.728854 4771 scope.go:117] "RemoveContainer" containerID="d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.739507 4771 generic.go:334] "Generic (PLEG): container finished" podID="2cb58c29-0f39-4260-85ce-9a93e3b90d99" containerID="f8a1dcc67e400ba67f2ef3eb32afd6f507f70cefc0023657d10303fcc00ecf81" exitCode=0 Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.739703 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fc64-account-create-c4fs5" event={"ID":"2cb58c29-0f39-4260-85ce-9a93e3b90d99","Type":"ContainerDied","Data":"f8a1dcc67e400ba67f2ef3eb32afd6f507f70cefc0023657d10303fcc00ecf81"} Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.740010 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fc64-account-create-c4fs5" event={"ID":"2cb58c29-0f39-4260-85ce-9a93e3b90d99","Type":"ContainerStarted","Data":"74afd284f0ffefd3ec35d1d4bd4b4bb6b6505ed078366c53f1006b3805c67120"} Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.747744 4771 generic.go:334] "Generic (PLEG): container finished" podID="4bbca829-8863-481a-8ff3-661dc1bc9726" containerID="251074125b8b3d0d61da59671183fc3b07b65f2ffedca9cf5ee0a3d134d74a7c" exitCode=0 Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.747943 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0ebb-account-create-sv52h" event={"ID":"4bbca829-8863-481a-8ff3-661dc1bc9726","Type":"ContainerDied","Data":"251074125b8b3d0d61da59671183fc3b07b65f2ffedca9cf5ee0a3d134d74a7c"} Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.753888 4771 generic.go:334] "Generic (PLEG): container finished" podID="8d3458e9-340f-43bf-a070-7e3ec1250176" containerID="de192baf53497d82961fd1234b70f275cb41f93639d15ee54a1f439dead69522" exitCode=0 Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.753990 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-053a-account-create-jt258" event={"ID":"8d3458e9-340f-43bf-a070-7e3ec1250176","Type":"ContainerDied","Data":"de192baf53497d82961fd1234b70f275cb41f93639d15ee54a1f439dead69522"} Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.754054 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-053a-account-create-jt258" event={"ID":"8d3458e9-340f-43bf-a070-7e3ec1250176","Type":"ContainerStarted","Data":"b2b1f9d60f9383e19dd0c47452a67ddb8609d3e0847dc8eaed3bee95d51e95bf"} Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.754819 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.754863 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.775467 4771 scope.go:117] "RemoveContainer" containerID="d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98" Oct 02 10:02:33 crc kubenswrapper[4771]: E1002 10:02:33.776035 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98\": container with ID starting with d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98 not found: ID does not exist" containerID="d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.776090 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98"} err="failed to get container status \"d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98\": rpc error: code = NotFound desc = could not find container \"d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98\": container with ID starting with d23e6578c1f2e83bee0ced97fca1ba9649d0eb309bcc7a0fe106fa153adb8f98 not found: ID does not exist" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.797939 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-config-data\") pod \"d7b40340-a806-490f-9842-a074ffba2e93\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.798035 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf9z8\" (UniqueName: \"kubernetes.io/projected/d7b40340-a806-490f-9842-a074ffba2e93-kube-api-access-tf9z8\") pod \"d7b40340-a806-490f-9842-a074ffba2e93\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.798103 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-config-data-custom\") pod \"d7b40340-a806-490f-9842-a074ffba2e93\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.798618 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-combined-ca-bundle\") pod \"d7b40340-a806-490f-9842-a074ffba2e93\" (UID: \"d7b40340-a806-490f-9842-a074ffba2e93\") " Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.807936 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d7b40340-a806-490f-9842-a074ffba2e93" (UID: "d7b40340-a806-490f-9842-a074ffba2e93"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.841411 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7b40340-a806-490f-9842-a074ffba2e93-kube-api-access-tf9z8" (OuterVolumeSpecName: "kube-api-access-tf9z8") pod "d7b40340-a806-490f-9842-a074ffba2e93" (UID: "d7b40340-a806-490f-9842-a074ffba2e93"). InnerVolumeSpecName "kube-api-access-tf9z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.848858 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7b40340-a806-490f-9842-a074ffba2e93" (UID: "d7b40340-a806-490f-9842-a074ffba2e93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.890780 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-config-data" (OuterVolumeSpecName: "config-data") pod "d7b40340-a806-490f-9842-a074ffba2e93" (UID: "d7b40340-a806-490f-9842-a074ffba2e93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.905204 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.905239 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.905248 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf9z8\" (UniqueName: \"kubernetes.io/projected/d7b40340-a806-490f-9842-a074ffba2e93-kube-api-access-tf9z8\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:33 crc kubenswrapper[4771]: I1002 10:02:33.905260 4771 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7b40340-a806-490f-9842-a074ffba2e93-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:34 crc kubenswrapper[4771]: I1002 10:02:34.723899 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 02 10:02:34 crc kubenswrapper[4771]: I1002 10:02:34.766850 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-649769f8cf-k8cph" Oct 02 10:02:34 crc kubenswrapper[4771]: I1002 10:02:34.824411 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-649769f8cf-k8cph"] Oct 02 10:02:34 crc kubenswrapper[4771]: I1002 10:02:34.843500 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-649769f8cf-k8cph"] Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.315215 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.315279 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.378268 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.379053 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.621470 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0ebb-account-create-sv52h" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.655314 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf7hm\" (UniqueName: \"kubernetes.io/projected/4bbca829-8863-481a-8ff3-661dc1bc9726-kube-api-access-vf7hm\") pod \"4bbca829-8863-481a-8ff3-661dc1bc9726\" (UID: \"4bbca829-8863-481a-8ff3-661dc1bc9726\") " Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.674392 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bbca829-8863-481a-8ff3-661dc1bc9726-kube-api-access-vf7hm" (OuterVolumeSpecName: "kube-api-access-vf7hm") pod "4bbca829-8863-481a-8ff3-661dc1bc9726" (UID: "4bbca829-8863-481a-8ff3-661dc1bc9726"). InnerVolumeSpecName "kube-api-access-vf7hm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.703497 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7b40340-a806-490f-9842-a074ffba2e93" path="/var/lib/kubelet/pods/d7b40340-a806-490f-9842-a074ffba2e93/volumes" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.760635 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf7hm\" (UniqueName: \"kubernetes.io/projected/4bbca829-8863-481a-8ff3-661dc1bc9726-kube-api-access-vf7hm\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.820888 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0ebb-account-create-sv52h" event={"ID":"4bbca829-8863-481a-8ff3-661dc1bc9726","Type":"ContainerDied","Data":"c4a1f386287aba9d2944dc5fe5fb03cbaf7eb4b2f211b430d31441904ae05ce1"} Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.820929 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4a1f386287aba9d2944dc5fe5fb03cbaf7eb4b2f211b430d31441904ae05ce1" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.821010 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0ebb-account-create-sv52h" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.840761 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-053a-account-create-jt258" event={"ID":"8d3458e9-340f-43bf-a070-7e3ec1250176","Type":"ContainerDied","Data":"b2b1f9d60f9383e19dd0c47452a67ddb8609d3e0847dc8eaed3bee95d51e95bf"} Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.840813 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2b1f9d60f9383e19dd0c47452a67ddb8609d3e0847dc8eaed3bee95d51e95bf" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.876835 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fc64-account-create-c4fs5" event={"ID":"2cb58c29-0f39-4260-85ce-9a93e3b90d99","Type":"ContainerDied","Data":"74afd284f0ffefd3ec35d1d4bd4b4bb6b6505ed078366c53f1006b3805c67120"} Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.876875 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74afd284f0ffefd3ec35d1d4bd4b4bb6b6505ed078366c53f1006b3805c67120" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.876897 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.877002 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.886438 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fc64-account-create-c4fs5" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.902958 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-053a-account-create-jt258" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.968363 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf2nt\" (UniqueName: \"kubernetes.io/projected/2cb58c29-0f39-4260-85ce-9a93e3b90d99-kube-api-access-sf2nt\") pod \"2cb58c29-0f39-4260-85ce-9a93e3b90d99\" (UID: \"2cb58c29-0f39-4260-85ce-9a93e3b90d99\") " Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.968823 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwzkr\" (UniqueName: \"kubernetes.io/projected/8d3458e9-340f-43bf-a070-7e3ec1250176-kube-api-access-wwzkr\") pod \"8d3458e9-340f-43bf-a070-7e3ec1250176\" (UID: \"8d3458e9-340f-43bf-a070-7e3ec1250176\") " Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.986078 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cb58c29-0f39-4260-85ce-9a93e3b90d99-kube-api-access-sf2nt" (OuterVolumeSpecName: "kube-api-access-sf2nt") pod "2cb58c29-0f39-4260-85ce-9a93e3b90d99" (UID: "2cb58c29-0f39-4260-85ce-9a93e3b90d99"). InnerVolumeSpecName "kube-api-access-sf2nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:35 crc kubenswrapper[4771]: I1002 10:02:35.986196 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d3458e9-340f-43bf-a070-7e3ec1250176-kube-api-access-wwzkr" (OuterVolumeSpecName: "kube-api-access-wwzkr") pod "8d3458e9-340f-43bf-a070-7e3ec1250176" (UID: "8d3458e9-340f-43bf-a070-7e3ec1250176"). InnerVolumeSpecName "kube-api-access-wwzkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:36 crc kubenswrapper[4771]: I1002 10:02:36.084719 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwzkr\" (UniqueName: \"kubernetes.io/projected/8d3458e9-340f-43bf-a070-7e3ec1250176-kube-api-access-wwzkr\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:36 crc kubenswrapper[4771]: I1002 10:02:36.085072 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf2nt\" (UniqueName: \"kubernetes.io/projected/2cb58c29-0f39-4260-85ce-9a93e3b90d99-kube-api-access-sf2nt\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:36 crc kubenswrapper[4771]: I1002 10:02:36.909475 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-053a-account-create-jt258" Oct 02 10:02:36 crc kubenswrapper[4771]: I1002 10:02:36.913252 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fc64-account-create-c4fs5" Oct 02 10:02:37 crc kubenswrapper[4771]: I1002 10:02:37.401607 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 10:02:37 crc kubenswrapper[4771]: I1002 10:02:37.402208 4771 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 10:02:37 crc kubenswrapper[4771]: I1002 10:02:37.460408 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 10:02:37 crc kubenswrapper[4771]: I1002 10:02:37.919418 4771 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 10:02:37 crc kubenswrapper[4771]: I1002 10:02:37.919468 4771 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 10:02:38 crc kubenswrapper[4771]: E1002 10:02:38.581032 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice/crio-8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f\": RecentStats: unable to find data in memory cache]" Oct 02 10:02:38 crc kubenswrapper[4771]: I1002 10:02:38.592673 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 10:02:38 crc kubenswrapper[4771]: I1002 10:02:38.612142 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.001429 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6f8xc"] Oct 02 10:02:42 crc kubenswrapper[4771]: E1002 10:02:42.002406 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbca829-8863-481a-8ff3-661dc1bc9726" containerName="mariadb-account-create" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.002420 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbca829-8863-481a-8ff3-661dc1bc9726" containerName="mariadb-account-create" Oct 02 10:02:42 crc kubenswrapper[4771]: E1002 10:02:42.002441 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b40340-a806-490f-9842-a074ffba2e93" containerName="heat-engine" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.002448 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b40340-a806-490f-9842-a074ffba2e93" containerName="heat-engine" Oct 02 10:02:42 crc kubenswrapper[4771]: E1002 10:02:42.002485 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d3458e9-340f-43bf-a070-7e3ec1250176" containerName="mariadb-account-create" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.002491 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d3458e9-340f-43bf-a070-7e3ec1250176" containerName="mariadb-account-create" Oct 02 10:02:42 crc kubenswrapper[4771]: E1002 10:02:42.002506 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cb58c29-0f39-4260-85ce-9a93e3b90d99" containerName="mariadb-account-create" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.002512 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cb58c29-0f39-4260-85ce-9a93e3b90d99" containerName="mariadb-account-create" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.002701 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7b40340-a806-490f-9842-a074ffba2e93" containerName="heat-engine" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.002731 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d3458e9-340f-43bf-a070-7e3ec1250176" containerName="mariadb-account-create" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.002741 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bbca829-8863-481a-8ff3-661dc1bc9726" containerName="mariadb-account-create" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.002762 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cb58c29-0f39-4260-85ce-9a93e3b90d99" containerName="mariadb-account-create" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.003555 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.007722 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-np9qq" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.008775 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.009299 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.074353 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-config-data\") pod \"nova-cell0-conductor-db-sync-6f8xc\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.074460 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6f8xc\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.074596 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kf8m\" (UniqueName: \"kubernetes.io/projected/d32f933a-a888-44bf-8028-19623221e336-kube-api-access-4kf8m\") pod \"nova-cell0-conductor-db-sync-6f8xc\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.074692 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-scripts\") pod \"nova-cell0-conductor-db-sync-6f8xc\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.097514 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6f8xc"] Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.146625 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.146689 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.176489 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kf8m\" (UniqueName: \"kubernetes.io/projected/d32f933a-a888-44bf-8028-19623221e336-kube-api-access-4kf8m\") pod \"nova-cell0-conductor-db-sync-6f8xc\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.176980 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-scripts\") pod \"nova-cell0-conductor-db-sync-6f8xc\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.178088 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-config-data\") pod \"nova-cell0-conductor-db-sync-6f8xc\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.178622 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6f8xc\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.196631 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-scripts\") pod \"nova-cell0-conductor-db-sync-6f8xc\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.197147 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6f8xc\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.200332 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kf8m\" (UniqueName: \"kubernetes.io/projected/d32f933a-a888-44bf-8028-19623221e336-kube-api-access-4kf8m\") pod \"nova-cell0-conductor-db-sync-6f8xc\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.203588 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-config-data\") pod \"nova-cell0-conductor-db-sync-6f8xc\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:42 crc kubenswrapper[4771]: I1002 10:02:42.394481 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:02:43 crc kubenswrapper[4771]: I1002 10:02:43.045611 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6f8xc"] Oct 02 10:02:43 crc kubenswrapper[4771]: W1002 10:02:43.084296 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd32f933a_a888_44bf_8028_19623221e336.slice/crio-390f41a23c7d7614d92f3bb0343339e996074eb5a3bd8768bfdc6fc683181e64 WatchSource:0}: Error finding container 390f41a23c7d7614d92f3bb0343339e996074eb5a3bd8768bfdc6fc683181e64: Status 404 returned error can't find the container with id 390f41a23c7d7614d92f3bb0343339e996074eb5a3bd8768bfdc6fc683181e64 Oct 02 10:02:44 crc kubenswrapper[4771]: I1002 10:02:44.036993 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6f8xc" event={"ID":"d32f933a-a888-44bf-8028-19623221e336","Type":"ContainerStarted","Data":"390f41a23c7d7614d92f3bb0343339e996074eb5a3bd8768bfdc6fc683181e64"} Oct 02 10:02:47 crc kubenswrapper[4771]: E1002 10:02:47.688833 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice/crio-8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice\": RecentStats: unable to find data in memory cache]" Oct 02 10:02:48 crc kubenswrapper[4771]: E1002 10:02:48.106556 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice/crio-8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice\": RecentStats: unable to find data in memory cache]" Oct 02 10:02:48 crc kubenswrapper[4771]: E1002 10:02:48.108079 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice/crio-8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice\": RecentStats: unable to find data in memory cache]" Oct 02 10:02:48 crc kubenswrapper[4771]: E1002 10:02:48.641918 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice/crio-8cb3b92089b314ed3c7290339afa0bf22a71969a3cb169aaf4183d602e39c24f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3184235b_9379_4b54_803d_dbb52582e06b.slice\": RecentStats: unable to find data in memory cache]" Oct 02 10:02:52 crc kubenswrapper[4771]: I1002 10:02:52.958423 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.066225 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-combined-ca-bundle\") pod \"3e1a0457-04c3-4dd6-8065-a04d208525c9\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.066525 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fnc5\" (UniqueName: \"kubernetes.io/projected/3e1a0457-04c3-4dd6-8065-a04d208525c9-kube-api-access-2fnc5\") pod \"3e1a0457-04c3-4dd6-8065-a04d208525c9\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.066606 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e1a0457-04c3-4dd6-8065-a04d208525c9-log-httpd\") pod \"3e1a0457-04c3-4dd6-8065-a04d208525c9\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.066687 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-scripts\") pod \"3e1a0457-04c3-4dd6-8065-a04d208525c9\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.066833 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e1a0457-04c3-4dd6-8065-a04d208525c9-run-httpd\") pod \"3e1a0457-04c3-4dd6-8065-a04d208525c9\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.066864 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-config-data\") pod \"3e1a0457-04c3-4dd6-8065-a04d208525c9\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.066963 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-sg-core-conf-yaml\") pod \"3e1a0457-04c3-4dd6-8065-a04d208525c9\" (UID: \"3e1a0457-04c3-4dd6-8065-a04d208525c9\") " Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.067120 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e1a0457-04c3-4dd6-8065-a04d208525c9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3e1a0457-04c3-4dd6-8065-a04d208525c9" (UID: "3e1a0457-04c3-4dd6-8065-a04d208525c9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.067512 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e1a0457-04c3-4dd6-8065-a04d208525c9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3e1a0457-04c3-4dd6-8065-a04d208525c9" (UID: "3e1a0457-04c3-4dd6-8065-a04d208525c9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.068721 4771 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e1a0457-04c3-4dd6-8065-a04d208525c9-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.068752 4771 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e1a0457-04c3-4dd6-8065-a04d208525c9-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.073494 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-scripts" (OuterVolumeSpecName: "scripts") pod "3e1a0457-04c3-4dd6-8065-a04d208525c9" (UID: "3e1a0457-04c3-4dd6-8065-a04d208525c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.073497 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e1a0457-04c3-4dd6-8065-a04d208525c9-kube-api-access-2fnc5" (OuterVolumeSpecName: "kube-api-access-2fnc5") pod "3e1a0457-04c3-4dd6-8065-a04d208525c9" (UID: "3e1a0457-04c3-4dd6-8065-a04d208525c9"). InnerVolumeSpecName "kube-api-access-2fnc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.122465 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3e1a0457-04c3-4dd6-8065-a04d208525c9" (UID: "3e1a0457-04c3-4dd6-8065-a04d208525c9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.174763 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fnc5\" (UniqueName: \"kubernetes.io/projected/3e1a0457-04c3-4dd6-8065-a04d208525c9-kube-api-access-2fnc5\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.174801 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.174814 4771 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.177366 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6f8xc" event={"ID":"d32f933a-a888-44bf-8028-19623221e336","Type":"ContainerStarted","Data":"557e4a57adaef22405d2d3b8213358bff7135decebaa687f4fa6039a1ad457f4"} Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.205202 4771 generic.go:334] "Generic (PLEG): container finished" podID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerID="9861341d50ff9c9b78d6835764f91d0fc5b0830106432d41bdeb9ba553d7a0b9" exitCode=137 Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.205257 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e1a0457-04c3-4dd6-8065-a04d208525c9","Type":"ContainerDied","Data":"9861341d50ff9c9b78d6835764f91d0fc5b0830106432d41bdeb9ba553d7a0b9"} Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.205287 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e1a0457-04c3-4dd6-8065-a04d208525c9","Type":"ContainerDied","Data":"d376784f90450d7039ab9198a34157220abc1f9e433a482d0d295104a844b3e9"} Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.205309 4771 scope.go:117] "RemoveContainer" containerID="9861341d50ff9c9b78d6835764f91d0fc5b0830106432d41bdeb9ba553d7a0b9" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.205477 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.212384 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e1a0457-04c3-4dd6-8065-a04d208525c9" (UID: "3e1a0457-04c3-4dd6-8065-a04d208525c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.234280 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-6f8xc" podStartSLOduration=2.922087758 podStartE2EDuration="12.23426079s" podCreationTimestamp="2025-10-02 10:02:41 +0000 UTC" firstStartedPulling="2025-10-02 10:02:43.090897388 +0000 UTC m=+1550.738582455" lastFinishedPulling="2025-10-02 10:02:52.40307042 +0000 UTC m=+1560.050755487" observedRunningTime="2025-10-02 10:02:53.230584906 +0000 UTC m=+1560.878269973" watchObservedRunningTime="2025-10-02 10:02:53.23426079 +0000 UTC m=+1560.881945857" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.260288 4771 scope.go:117] "RemoveContainer" containerID="aebc0b89a5d5628ec59f5be0429060dd4ca06f624db3f883833edf6a3e93229b" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.271363 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-config-data" (OuterVolumeSpecName: "config-data") pod "3e1a0457-04c3-4dd6-8065-a04d208525c9" (UID: "3e1a0457-04c3-4dd6-8065-a04d208525c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.279885 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.279921 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1a0457-04c3-4dd6-8065-a04d208525c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.306338 4771 scope.go:117] "RemoveContainer" containerID="a3912405d27c0f84507547d3e57a9a434827c85fe0ac745c017c03df7dba93fb" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.346336 4771 scope.go:117] "RemoveContainer" containerID="fbe281d1e373fb76203172136537c17314552d8df3d2fa6c50de186d8b9cbd15" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.482283 4771 scope.go:117] "RemoveContainer" containerID="9861341d50ff9c9b78d6835764f91d0fc5b0830106432d41bdeb9ba553d7a0b9" Oct 02 10:02:53 crc kubenswrapper[4771]: E1002 10:02:53.492440 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9861341d50ff9c9b78d6835764f91d0fc5b0830106432d41bdeb9ba553d7a0b9\": container with ID starting with 9861341d50ff9c9b78d6835764f91d0fc5b0830106432d41bdeb9ba553d7a0b9 not found: ID does not exist" containerID="9861341d50ff9c9b78d6835764f91d0fc5b0830106432d41bdeb9ba553d7a0b9" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.492496 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9861341d50ff9c9b78d6835764f91d0fc5b0830106432d41bdeb9ba553d7a0b9"} err="failed to get container status \"9861341d50ff9c9b78d6835764f91d0fc5b0830106432d41bdeb9ba553d7a0b9\": rpc error: code = NotFound desc = could not find container \"9861341d50ff9c9b78d6835764f91d0fc5b0830106432d41bdeb9ba553d7a0b9\": container with ID starting with 9861341d50ff9c9b78d6835764f91d0fc5b0830106432d41bdeb9ba553d7a0b9 not found: ID does not exist" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.492526 4771 scope.go:117] "RemoveContainer" containerID="aebc0b89a5d5628ec59f5be0429060dd4ca06f624db3f883833edf6a3e93229b" Oct 02 10:02:53 crc kubenswrapper[4771]: E1002 10:02:53.496777 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aebc0b89a5d5628ec59f5be0429060dd4ca06f624db3f883833edf6a3e93229b\": container with ID starting with aebc0b89a5d5628ec59f5be0429060dd4ca06f624db3f883833edf6a3e93229b not found: ID does not exist" containerID="aebc0b89a5d5628ec59f5be0429060dd4ca06f624db3f883833edf6a3e93229b" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.496821 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aebc0b89a5d5628ec59f5be0429060dd4ca06f624db3f883833edf6a3e93229b"} err="failed to get container status \"aebc0b89a5d5628ec59f5be0429060dd4ca06f624db3f883833edf6a3e93229b\": rpc error: code = NotFound desc = could not find container \"aebc0b89a5d5628ec59f5be0429060dd4ca06f624db3f883833edf6a3e93229b\": container with ID starting with aebc0b89a5d5628ec59f5be0429060dd4ca06f624db3f883833edf6a3e93229b not found: ID does not exist" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.496848 4771 scope.go:117] "RemoveContainer" containerID="a3912405d27c0f84507547d3e57a9a434827c85fe0ac745c017c03df7dba93fb" Oct 02 10:02:53 crc kubenswrapper[4771]: E1002 10:02:53.497770 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3912405d27c0f84507547d3e57a9a434827c85fe0ac745c017c03df7dba93fb\": container with ID starting with a3912405d27c0f84507547d3e57a9a434827c85fe0ac745c017c03df7dba93fb not found: ID does not exist" containerID="a3912405d27c0f84507547d3e57a9a434827c85fe0ac745c017c03df7dba93fb" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.497819 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3912405d27c0f84507547d3e57a9a434827c85fe0ac745c017c03df7dba93fb"} err="failed to get container status \"a3912405d27c0f84507547d3e57a9a434827c85fe0ac745c017c03df7dba93fb\": rpc error: code = NotFound desc = could not find container \"a3912405d27c0f84507547d3e57a9a434827c85fe0ac745c017c03df7dba93fb\": container with ID starting with a3912405d27c0f84507547d3e57a9a434827c85fe0ac745c017c03df7dba93fb not found: ID does not exist" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.497852 4771 scope.go:117] "RemoveContainer" containerID="fbe281d1e373fb76203172136537c17314552d8df3d2fa6c50de186d8b9cbd15" Oct 02 10:02:53 crc kubenswrapper[4771]: E1002 10:02:53.498317 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbe281d1e373fb76203172136537c17314552d8df3d2fa6c50de186d8b9cbd15\": container with ID starting with fbe281d1e373fb76203172136537c17314552d8df3d2fa6c50de186d8b9cbd15 not found: ID does not exist" containerID="fbe281d1e373fb76203172136537c17314552d8df3d2fa6c50de186d8b9cbd15" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.498341 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbe281d1e373fb76203172136537c17314552d8df3d2fa6c50de186d8b9cbd15"} err="failed to get container status \"fbe281d1e373fb76203172136537c17314552d8df3d2fa6c50de186d8b9cbd15\": rpc error: code = NotFound desc = could not find container \"fbe281d1e373fb76203172136537c17314552d8df3d2fa6c50de186d8b9cbd15\": container with ID starting with fbe281d1e373fb76203172136537c17314552d8df3d2fa6c50de186d8b9cbd15 not found: ID does not exist" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.568201 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.588699 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.612203 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:02:53 crc kubenswrapper[4771]: E1002 10:02:53.612904 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="ceilometer-central-agent" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.612933 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="ceilometer-central-agent" Oct 02 10:02:53 crc kubenswrapper[4771]: E1002 10:02:53.612988 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="proxy-httpd" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.613000 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="proxy-httpd" Oct 02 10:02:53 crc kubenswrapper[4771]: E1002 10:02:53.613014 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="sg-core" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.613022 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="sg-core" Oct 02 10:02:53 crc kubenswrapper[4771]: E1002 10:02:53.613043 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="ceilometer-notification-agent" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.613052 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="ceilometer-notification-agent" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.613367 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="sg-core" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.613392 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="proxy-httpd" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.613406 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="ceilometer-notification-agent" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.613418 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" containerName="ceilometer-central-agent" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.615725 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.619216 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.621742 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.639326 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.705420 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e1a0457-04c3-4dd6-8065-a04d208525c9" path="/var/lib/kubelet/pods/3e1a0457-04c3-4dd6-8065-a04d208525c9/volumes" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.795101 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.795602 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnmmf\" (UniqueName: \"kubernetes.io/projected/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-kube-api-access-xnmmf\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.795682 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-log-httpd\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.795789 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.796261 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-scripts\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.796359 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-config-data\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.796441 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-run-httpd\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.898925 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-scripts\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.898991 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-config-data\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.899041 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-run-httpd\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.899088 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.899142 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnmmf\" (UniqueName: \"kubernetes.io/projected/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-kube-api-access-xnmmf\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.899188 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-log-httpd\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.899209 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.900553 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-run-httpd\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.901253 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-log-httpd\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.904376 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-scripts\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.904602 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-config-data\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.904970 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.906392 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.919254 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnmmf\" (UniqueName: \"kubernetes.io/projected/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-kube-api-access-xnmmf\") pod \"ceilometer-0\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " pod="openstack/ceilometer-0" Oct 02 10:02:53 crc kubenswrapper[4771]: I1002 10:02:53.937196 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:02:54 crc kubenswrapper[4771]: I1002 10:02:54.000506 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:02:54 crc kubenswrapper[4771]: I1002 10:02:54.520287 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:02:55 crc kubenswrapper[4771]: I1002 10:02:55.243692 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2","Type":"ContainerStarted","Data":"9751d00742c1afb0054a21bb291bfb1d4995eda0b3c125814973d14919b18398"} Oct 02 10:02:56 crc kubenswrapper[4771]: I1002 10:02:56.258834 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2","Type":"ContainerStarted","Data":"8e0e6c546d868fb516f528a3debf128acb06fb886c1937d41ad3cf733412532d"} Oct 02 10:02:57 crc kubenswrapper[4771]: I1002 10:02:57.275703 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2","Type":"ContainerStarted","Data":"6150b5710858c3bf50cf2ff0aa0ff8d70707cbfdad4049c895b9d7c71e8e32e5"} Oct 02 10:02:58 crc kubenswrapper[4771]: I1002 10:02:58.293229 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2","Type":"ContainerStarted","Data":"567d20c4f471cbcee84b90a9b94e975629fc44f4e074aa88e18535cdb9b3d631"} Oct 02 10:03:02 crc kubenswrapper[4771]: I1002 10:03:02.346257 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2","Type":"ContainerStarted","Data":"583237c0608681d5f34efbcc783349fb21204003c757ec54f512c549181dc0de"} Oct 02 10:03:02 crc kubenswrapper[4771]: I1002 10:03:02.348328 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:03:02 crc kubenswrapper[4771]: I1002 10:03:02.346558 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="sg-core" containerID="cri-o://567d20c4f471cbcee84b90a9b94e975629fc44f4e074aa88e18535cdb9b3d631" gracePeriod=30 Oct 02 10:03:02 crc kubenswrapper[4771]: I1002 10:03:02.346612 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="ceilometer-notification-agent" containerID="cri-o://6150b5710858c3bf50cf2ff0aa0ff8d70707cbfdad4049c895b9d7c71e8e32e5" gracePeriod=30 Oct 02 10:03:02 crc kubenswrapper[4771]: I1002 10:03:02.346634 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="proxy-httpd" containerID="cri-o://583237c0608681d5f34efbcc783349fb21204003c757ec54f512c549181dc0de" gracePeriod=30 Oct 02 10:03:02 crc kubenswrapper[4771]: I1002 10:03:02.346470 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="ceilometer-central-agent" containerID="cri-o://8e0e6c546d868fb516f528a3debf128acb06fb886c1937d41ad3cf733412532d" gracePeriod=30 Oct 02 10:03:02 crc kubenswrapper[4771]: I1002 10:03:02.377577 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.123505223 podStartE2EDuration="9.377557243s" podCreationTimestamp="2025-10-02 10:02:53 +0000 UTC" firstStartedPulling="2025-10-02 10:02:54.509649894 +0000 UTC m=+1562.157334961" lastFinishedPulling="2025-10-02 10:03:00.763701914 +0000 UTC m=+1568.411386981" observedRunningTime="2025-10-02 10:03:02.370272746 +0000 UTC m=+1570.017957833" watchObservedRunningTime="2025-10-02 10:03:02.377557243 +0000 UTC m=+1570.025242310" Oct 02 10:03:03 crc kubenswrapper[4771]: I1002 10:03:03.372338 4771 generic.go:334] "Generic (PLEG): container finished" podID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerID="567d20c4f471cbcee84b90a9b94e975629fc44f4e074aa88e18535cdb9b3d631" exitCode=2 Oct 02 10:03:03 crc kubenswrapper[4771]: I1002 10:03:03.372719 4771 generic.go:334] "Generic (PLEG): container finished" podID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerID="6150b5710858c3bf50cf2ff0aa0ff8d70707cbfdad4049c895b9d7c71e8e32e5" exitCode=0 Oct 02 10:03:03 crc kubenswrapper[4771]: I1002 10:03:03.372749 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2","Type":"ContainerDied","Data":"567d20c4f471cbcee84b90a9b94e975629fc44f4e074aa88e18535cdb9b3d631"} Oct 02 10:03:03 crc kubenswrapper[4771]: I1002 10:03:03.372787 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2","Type":"ContainerDied","Data":"6150b5710858c3bf50cf2ff0aa0ff8d70707cbfdad4049c895b9d7c71e8e32e5"} Oct 02 10:03:04 crc kubenswrapper[4771]: I1002 10:03:04.387343 4771 generic.go:334] "Generic (PLEG): container finished" podID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerID="583237c0608681d5f34efbcc783349fb21204003c757ec54f512c549181dc0de" exitCode=0 Oct 02 10:03:04 crc kubenswrapper[4771]: I1002 10:03:04.387407 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2","Type":"ContainerDied","Data":"583237c0608681d5f34efbcc783349fb21204003c757ec54f512c549181dc0de"} Oct 02 10:03:12 crc kubenswrapper[4771]: I1002 10:03:12.146191 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:03:12 crc kubenswrapper[4771]: I1002 10:03:12.148070 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:03:12 crc kubenswrapper[4771]: I1002 10:03:12.148260 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 10:03:12 crc kubenswrapper[4771]: I1002 10:03:12.149360 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:03:12 crc kubenswrapper[4771]: I1002 10:03:12.149567 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" gracePeriod=600 Oct 02 10:03:12 crc kubenswrapper[4771]: I1002 10:03:12.485482 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" exitCode=0 Oct 02 10:03:12 crc kubenswrapper[4771]: I1002 10:03:12.485881 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772"} Oct 02 10:03:12 crc kubenswrapper[4771]: I1002 10:03:12.485920 4771 scope.go:117] "RemoveContainer" containerID="405cf9bc15cbb64122966859b742b139fd0128a50163288165936c501ca2e3e6" Oct 02 10:03:12 crc kubenswrapper[4771]: E1002 10:03:12.787591 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:03:13 crc kubenswrapper[4771]: I1002 10:03:13.533846 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:03:13 crc kubenswrapper[4771]: E1002 10:03:13.535069 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:03:13 crc kubenswrapper[4771]: I1002 10:03:13.548358 4771 generic.go:334] "Generic (PLEG): container finished" podID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerID="8e0e6c546d868fb516f528a3debf128acb06fb886c1937d41ad3cf733412532d" exitCode=0 Oct 02 10:03:13 crc kubenswrapper[4771]: I1002 10:03:13.548425 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2","Type":"ContainerDied","Data":"8e0e6c546d868fb516f528a3debf128acb06fb886c1937d41ad3cf733412532d"} Oct 02 10:03:13 crc kubenswrapper[4771]: I1002 10:03:13.890421 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.017027 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-sg-core-conf-yaml\") pod \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.017114 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-log-httpd\") pod \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.017195 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-scripts\") pod \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.017273 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-config-data\") pod \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.017304 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-run-httpd\") pod \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.017351 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnmmf\" (UniqueName: \"kubernetes.io/projected/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-kube-api-access-xnmmf\") pod \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.017367 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-combined-ca-bundle\") pod \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\" (UID: \"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2\") " Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.019109 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" (UID: "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.021659 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" (UID: "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.031827 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-kube-api-access-xnmmf" (OuterVolumeSpecName: "kube-api-access-xnmmf") pod "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" (UID: "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2"). InnerVolumeSpecName "kube-api-access-xnmmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.033352 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-scripts" (OuterVolumeSpecName: "scripts") pod "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" (UID: "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.065330 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" (UID: "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.115943 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" (UID: "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.119816 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnmmf\" (UniqueName: \"kubernetes.io/projected/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-kube-api-access-xnmmf\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.119843 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.119851 4771 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.119861 4771 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.119871 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.119879 4771 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.172366 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-config-data" (OuterVolumeSpecName: "config-data") pod "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" (UID: "8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.222100 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.565462 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.565448 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2","Type":"ContainerDied","Data":"9751d00742c1afb0054a21bb291bfb1d4995eda0b3c125814973d14919b18398"} Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.565597 4771 scope.go:117] "RemoveContainer" containerID="583237c0608681d5f34efbcc783349fb21204003c757ec54f512c549181dc0de" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.567183 4771 generic.go:334] "Generic (PLEG): container finished" podID="d32f933a-a888-44bf-8028-19623221e336" containerID="557e4a57adaef22405d2d3b8213358bff7135decebaa687f4fa6039a1ad457f4" exitCode=0 Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.567232 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6f8xc" event={"ID":"d32f933a-a888-44bf-8028-19623221e336","Type":"ContainerDied","Data":"557e4a57adaef22405d2d3b8213358bff7135decebaa687f4fa6039a1ad457f4"} Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.601732 4771 scope.go:117] "RemoveContainer" containerID="567d20c4f471cbcee84b90a9b94e975629fc44f4e074aa88e18535cdb9b3d631" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.615472 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.629348 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.640498 4771 scope.go:117] "RemoveContainer" containerID="6150b5710858c3bf50cf2ff0aa0ff8d70707cbfdad4049c895b9d7c71e8e32e5" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.645686 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:03:14 crc kubenswrapper[4771]: E1002 10:03:14.646207 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="ceilometer-notification-agent" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.646224 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="ceilometer-notification-agent" Oct 02 10:03:14 crc kubenswrapper[4771]: E1002 10:03:14.646246 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="ceilometer-central-agent" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.646254 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="ceilometer-central-agent" Oct 02 10:03:14 crc kubenswrapper[4771]: E1002 10:03:14.646297 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="sg-core" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.646303 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="sg-core" Oct 02 10:03:14 crc kubenswrapper[4771]: E1002 10:03:14.646321 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="proxy-httpd" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.646327 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="proxy-httpd" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.646542 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="sg-core" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.646564 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="ceilometer-notification-agent" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.646579 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="proxy-httpd" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.646598 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" containerName="ceilometer-central-agent" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.651686 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.655047 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.655461 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.657859 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.686229 4771 scope.go:117] "RemoveContainer" containerID="8e0e6c546d868fb516f528a3debf128acb06fb886c1937d41ad3cf733412532d" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.843300 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.844077 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-scripts\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.844251 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmlpr\" (UniqueName: \"kubernetes.io/projected/9158fae1-e129-47d9-af54-2105c7d79e41-kube-api-access-cmlpr\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.844421 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-config-data\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.844436 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9158fae1-e129-47d9-af54-2105c7d79e41-log-httpd\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.844498 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.844609 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9158fae1-e129-47d9-af54-2105c7d79e41-run-httpd\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.946806 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-scripts\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.946942 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmlpr\" (UniqueName: \"kubernetes.io/projected/9158fae1-e129-47d9-af54-2105c7d79e41-kube-api-access-cmlpr\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.947036 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-config-data\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.947057 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9158fae1-e129-47d9-af54-2105c7d79e41-log-httpd\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.947104 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.947179 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9158fae1-e129-47d9-af54-2105c7d79e41-run-httpd\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.947246 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.947687 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9158fae1-e129-47d9-af54-2105c7d79e41-log-httpd\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.948084 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9158fae1-e129-47d9-af54-2105c7d79e41-run-httpd\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.951506 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-scripts\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.951813 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.952440 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-config-data\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.961792 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:14 crc kubenswrapper[4771]: I1002 10:03:14.977073 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmlpr\" (UniqueName: \"kubernetes.io/projected/9158fae1-e129-47d9-af54-2105c7d79e41-kube-api-access-cmlpr\") pod \"ceilometer-0\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " pod="openstack/ceilometer-0" Oct 02 10:03:15 crc kubenswrapper[4771]: I1002 10:03:15.083452 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:03:15 crc kubenswrapper[4771]: I1002 10:03:15.623499 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:03:15 crc kubenswrapper[4771]: W1002 10:03:15.624522 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-7f5f125117d1b1388a8503a71ec4c602610220b192068f6624d985dd54031912 WatchSource:0}: Error finding container 7f5f125117d1b1388a8503a71ec4c602610220b192068f6624d985dd54031912: Status 404 returned error can't find the container with id 7f5f125117d1b1388a8503a71ec4c602610220b192068f6624d985dd54031912 Oct 02 10:03:15 crc kubenswrapper[4771]: I1002 10:03:15.709824 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2" path="/var/lib/kubelet/pods/8f2fe919-f0a9-4bfc-ae09-6bff66fb20f2/volumes" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.172094 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.299668 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-combined-ca-bundle\") pod \"d32f933a-a888-44bf-8028-19623221e336\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.299782 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-config-data\") pod \"d32f933a-a888-44bf-8028-19623221e336\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.299811 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-scripts\") pod \"d32f933a-a888-44bf-8028-19623221e336\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.299938 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kf8m\" (UniqueName: \"kubernetes.io/projected/d32f933a-a888-44bf-8028-19623221e336-kube-api-access-4kf8m\") pod \"d32f933a-a888-44bf-8028-19623221e336\" (UID: \"d32f933a-a888-44bf-8028-19623221e336\") " Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.305257 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d32f933a-a888-44bf-8028-19623221e336-kube-api-access-4kf8m" (OuterVolumeSpecName: "kube-api-access-4kf8m") pod "d32f933a-a888-44bf-8028-19623221e336" (UID: "d32f933a-a888-44bf-8028-19623221e336"). InnerVolumeSpecName "kube-api-access-4kf8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.305589 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-scripts" (OuterVolumeSpecName: "scripts") pod "d32f933a-a888-44bf-8028-19623221e336" (UID: "d32f933a-a888-44bf-8028-19623221e336"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.343871 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-config-data" (OuterVolumeSpecName: "config-data") pod "d32f933a-a888-44bf-8028-19623221e336" (UID: "d32f933a-a888-44bf-8028-19623221e336"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.347942 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d32f933a-a888-44bf-8028-19623221e336" (UID: "d32f933a-a888-44bf-8028-19623221e336"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.404600 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.404666 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.404681 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32f933a-a888-44bf-8028-19623221e336-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.404696 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kf8m\" (UniqueName: \"kubernetes.io/projected/d32f933a-a888-44bf-8028-19623221e336-kube-api-access-4kf8m\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.594587 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6f8xc" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.594589 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6f8xc" event={"ID":"d32f933a-a888-44bf-8028-19623221e336","Type":"ContainerDied","Data":"390f41a23c7d7614d92f3bb0343339e996074eb5a3bd8768bfdc6fc683181e64"} Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.594739 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="390f41a23c7d7614d92f3bb0343339e996074eb5a3bd8768bfdc6fc683181e64" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.596472 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9158fae1-e129-47d9-af54-2105c7d79e41","Type":"ContainerStarted","Data":"97fdbd831c00862cda41bbf77db00d97cea8c8d22204850b73ed21a0c5975289"} Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.596537 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9158fae1-e129-47d9-af54-2105c7d79e41","Type":"ContainerStarted","Data":"7f5f125117d1b1388a8503a71ec4c602610220b192068f6624d985dd54031912"} Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.726846 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 10:03:16 crc kubenswrapper[4771]: E1002 10:03:16.727409 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d32f933a-a888-44bf-8028-19623221e336" containerName="nova-cell0-conductor-db-sync" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.727425 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d32f933a-a888-44bf-8028-19623221e336" containerName="nova-cell0-conductor-db-sync" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.735307 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="d32f933a-a888-44bf-8028-19623221e336" containerName="nova-cell0-conductor-db-sync" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.736304 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.740747 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-np9qq" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.744163 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.754336 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.820680 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmr8x\" (UniqueName: \"kubernetes.io/projected/61c5ccc6-c403-431b-888a-71c9d79c702a-kube-api-access-gmr8x\") pod \"nova-cell0-conductor-0\" (UID: \"61c5ccc6-c403-431b-888a-71c9d79c702a\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.820858 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61c5ccc6-c403-431b-888a-71c9d79c702a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"61c5ccc6-c403-431b-888a-71c9d79c702a\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.820905 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61c5ccc6-c403-431b-888a-71c9d79c702a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"61c5ccc6-c403-431b-888a-71c9d79c702a\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.925497 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61c5ccc6-c403-431b-888a-71c9d79c702a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"61c5ccc6-c403-431b-888a-71c9d79c702a\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.925598 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61c5ccc6-c403-431b-888a-71c9d79c702a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"61c5ccc6-c403-431b-888a-71c9d79c702a\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.925859 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmr8x\" (UniqueName: \"kubernetes.io/projected/61c5ccc6-c403-431b-888a-71c9d79c702a-kube-api-access-gmr8x\") pod \"nova-cell0-conductor-0\" (UID: \"61c5ccc6-c403-431b-888a-71c9d79c702a\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.932898 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61c5ccc6-c403-431b-888a-71c9d79c702a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"61c5ccc6-c403-431b-888a-71c9d79c702a\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.933960 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61c5ccc6-c403-431b-888a-71c9d79c702a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"61c5ccc6-c403-431b-888a-71c9d79c702a\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:03:16 crc kubenswrapper[4771]: I1002 10:03:16.951802 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmr8x\" (UniqueName: \"kubernetes.io/projected/61c5ccc6-c403-431b-888a-71c9d79c702a-kube-api-access-gmr8x\") pod \"nova-cell0-conductor-0\" (UID: \"61c5ccc6-c403-431b-888a-71c9d79c702a\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:03:17 crc kubenswrapper[4771]: I1002 10:03:17.100325 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 10:03:17 crc kubenswrapper[4771]: I1002 10:03:17.612517 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9158fae1-e129-47d9-af54-2105c7d79e41","Type":"ContainerStarted","Data":"9467df9eaffc2ba3c84b6bce0d755e29d56809a3fee62731a1e3e535a9af9035"} Oct 02 10:03:17 crc kubenswrapper[4771]: I1002 10:03:17.653981 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 10:03:18 crc kubenswrapper[4771]: I1002 10:03:18.439394 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-bgtxq"] Oct 02 10:03:18 crc kubenswrapper[4771]: I1002 10:03:18.444109 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-bgtxq" Oct 02 10:03:18 crc kubenswrapper[4771]: I1002 10:03:18.484788 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-bgtxq"] Oct 02 10:03:18 crc kubenswrapper[4771]: I1002 10:03:18.587642 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s486p\" (UniqueName: \"kubernetes.io/projected/8d02770b-766e-488b-a144-dd9d8ecc0653-kube-api-access-s486p\") pod \"aodh-db-create-bgtxq\" (UID: \"8d02770b-766e-488b-a144-dd9d8ecc0653\") " pod="openstack/aodh-db-create-bgtxq" Oct 02 10:03:18 crc kubenswrapper[4771]: I1002 10:03:18.625580 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"61c5ccc6-c403-431b-888a-71c9d79c702a","Type":"ContainerStarted","Data":"1f78f4d578118fa0d2f2906084e701ebadda634bd86db03756741af31e514275"} Oct 02 10:03:18 crc kubenswrapper[4771]: I1002 10:03:18.626190 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 02 10:03:18 crc kubenswrapper[4771]: I1002 10:03:18.626245 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"61c5ccc6-c403-431b-888a-71c9d79c702a","Type":"ContainerStarted","Data":"564b8d5cb772c584d972dbeec2c8944f8770cb98c6a80f3a48241d30709c9f9c"} Oct 02 10:03:18 crc kubenswrapper[4771]: I1002 10:03:18.627997 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9158fae1-e129-47d9-af54-2105c7d79e41","Type":"ContainerStarted","Data":"da6a2166316f9f1ead80c45c13320c5f8cdf387ddf0ee787ebba09e521d4cf4e"} Oct 02 10:03:18 crc kubenswrapper[4771]: I1002 10:03:18.651300 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.651277711 podStartE2EDuration="2.651277711s" podCreationTimestamp="2025-10-02 10:03:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:03:18.646005091 +0000 UTC m=+1586.293690158" watchObservedRunningTime="2025-10-02 10:03:18.651277711 +0000 UTC m=+1586.298962768" Oct 02 10:03:18 crc kubenswrapper[4771]: I1002 10:03:18.690658 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s486p\" (UniqueName: \"kubernetes.io/projected/8d02770b-766e-488b-a144-dd9d8ecc0653-kube-api-access-s486p\") pod \"aodh-db-create-bgtxq\" (UID: \"8d02770b-766e-488b-a144-dd9d8ecc0653\") " pod="openstack/aodh-db-create-bgtxq" Oct 02 10:03:18 crc kubenswrapper[4771]: I1002 10:03:18.711148 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s486p\" (UniqueName: \"kubernetes.io/projected/8d02770b-766e-488b-a144-dd9d8ecc0653-kube-api-access-s486p\") pod \"aodh-db-create-bgtxq\" (UID: \"8d02770b-766e-488b-a144-dd9d8ecc0653\") " pod="openstack/aodh-db-create-bgtxq" Oct 02 10:03:18 crc kubenswrapper[4771]: I1002 10:03:18.774063 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-bgtxq" Oct 02 10:03:19 crc kubenswrapper[4771]: I1002 10:03:19.367246 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-bgtxq"] Oct 02 10:03:19 crc kubenswrapper[4771]: W1002 10:03:19.394944 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d02770b_766e_488b_a144_dd9d8ecc0653.slice/crio-3cb73ba31a2f578515473bc77af95dc8d1ff1565d8ddebd57e1c61d438232dfc WatchSource:0}: Error finding container 3cb73ba31a2f578515473bc77af95dc8d1ff1565d8ddebd57e1c61d438232dfc: Status 404 returned error can't find the container with id 3cb73ba31a2f578515473bc77af95dc8d1ff1565d8ddebd57e1c61d438232dfc Oct 02 10:03:19 crc kubenswrapper[4771]: I1002 10:03:19.650028 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9158fae1-e129-47d9-af54-2105c7d79e41","Type":"ContainerStarted","Data":"d708a48848c28ecb797ca149a414b0ed2e22f15e3d4c604bdf2da2ca2b780207"} Oct 02 10:03:19 crc kubenswrapper[4771]: I1002 10:03:19.655236 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-bgtxq" event={"ID":"8d02770b-766e-488b-a144-dd9d8ecc0653","Type":"ContainerStarted","Data":"3cb73ba31a2f578515473bc77af95dc8d1ff1565d8ddebd57e1c61d438232dfc"} Oct 02 10:03:19 crc kubenswrapper[4771]: I1002 10:03:19.688458 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.216567886 podStartE2EDuration="5.688440319s" podCreationTimestamp="2025-10-02 10:03:14 +0000 UTC" firstStartedPulling="2025-10-02 10:03:15.627825942 +0000 UTC m=+1583.275511009" lastFinishedPulling="2025-10-02 10:03:19.099698365 +0000 UTC m=+1586.747383442" observedRunningTime="2025-10-02 10:03:19.686435292 +0000 UTC m=+1587.334120359" watchObservedRunningTime="2025-10-02 10:03:19.688440319 +0000 UTC m=+1587.336125386" Oct 02 10:03:20 crc kubenswrapper[4771]: I1002 10:03:20.704782 4771 generic.go:334] "Generic (PLEG): container finished" podID="8d02770b-766e-488b-a144-dd9d8ecc0653" containerID="cbfa7a7774deaaf3dfdb7596c8b2a9c83d804e29ba043e5befcae2aa740057db" exitCode=0 Oct 02 10:03:20 crc kubenswrapper[4771]: I1002 10:03:20.705504 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-bgtxq" event={"ID":"8d02770b-766e-488b-a144-dd9d8ecc0653","Type":"ContainerDied","Data":"cbfa7a7774deaaf3dfdb7596c8b2a9c83d804e29ba043e5befcae2aa740057db"} Oct 02 10:03:20 crc kubenswrapper[4771]: I1002 10:03:20.706297 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:03:22 crc kubenswrapper[4771]: I1002 10:03:22.317920 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-bgtxq" Oct 02 10:03:22 crc kubenswrapper[4771]: I1002 10:03:22.516811 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s486p\" (UniqueName: \"kubernetes.io/projected/8d02770b-766e-488b-a144-dd9d8ecc0653-kube-api-access-s486p\") pod \"8d02770b-766e-488b-a144-dd9d8ecc0653\" (UID: \"8d02770b-766e-488b-a144-dd9d8ecc0653\") " Oct 02 10:03:22 crc kubenswrapper[4771]: I1002 10:03:22.524508 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d02770b-766e-488b-a144-dd9d8ecc0653-kube-api-access-s486p" (OuterVolumeSpecName: "kube-api-access-s486p") pod "8d02770b-766e-488b-a144-dd9d8ecc0653" (UID: "8d02770b-766e-488b-a144-dd9d8ecc0653"). InnerVolumeSpecName "kube-api-access-s486p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:22 crc kubenswrapper[4771]: I1002 10:03:22.620968 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s486p\" (UniqueName: \"kubernetes.io/projected/8d02770b-766e-488b-a144-dd9d8ecc0653-kube-api-access-s486p\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:22 crc kubenswrapper[4771]: I1002 10:03:22.736931 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-bgtxq" event={"ID":"8d02770b-766e-488b-a144-dd9d8ecc0653","Type":"ContainerDied","Data":"3cb73ba31a2f578515473bc77af95dc8d1ff1565d8ddebd57e1c61d438232dfc"} Oct 02 10:03:22 crc kubenswrapper[4771]: I1002 10:03:22.737354 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cb73ba31a2f578515473bc77af95dc8d1ff1565d8ddebd57e1c61d438232dfc" Oct 02 10:03:22 crc kubenswrapper[4771]: I1002 10:03:22.737182 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-bgtxq" Oct 02 10:03:25 crc kubenswrapper[4771]: I1002 10:03:25.681878 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:03:25 crc kubenswrapper[4771]: E1002 10:03:25.682944 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.140173 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.623054 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-mzcxs"] Oct 02 10:03:27 crc kubenswrapper[4771]: E1002 10:03:27.624825 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d02770b-766e-488b-a144-dd9d8ecc0653" containerName="mariadb-database-create" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.624851 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d02770b-766e-488b-a144-dd9d8ecc0653" containerName="mariadb-database-create" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.625220 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d02770b-766e-488b-a144-dd9d8ecc0653" containerName="mariadb-database-create" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.626474 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.641869 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-mzcxs"] Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.647221 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.647239 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.758044 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-scripts\") pod \"nova-cell0-cell-mapping-mzcxs\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.758145 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mzcxs\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.758251 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tff6t\" (UniqueName: \"kubernetes.io/projected/913531c6-8fa3-4c33-87a4-22f18bb251b8-kube-api-access-tff6t\") pod \"nova-cell0-cell-mapping-mzcxs\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.758315 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-config-data\") pod \"nova-cell0-cell-mapping-mzcxs\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.844760 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.846983 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.852694 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.861430 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-scripts\") pod \"nova-cell0-cell-mapping-mzcxs\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.861494 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mzcxs\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.861611 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tff6t\" (UniqueName: \"kubernetes.io/projected/913531c6-8fa3-4c33-87a4-22f18bb251b8-kube-api-access-tff6t\") pod \"nova-cell0-cell-mapping-mzcxs\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.861717 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-config-data\") pod \"nova-cell0-cell-mapping-mzcxs\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.872360 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mzcxs\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.878235 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-scripts\") pod \"nova-cell0-cell-mapping-mzcxs\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.883504 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.898804 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.901405 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.909884 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-config-data\") pod \"nova-cell0-cell-mapping-mzcxs\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.910183 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.938509 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tff6t\" (UniqueName: \"kubernetes.io/projected/913531c6-8fa3-4c33-87a4-22f18bb251b8-kube-api-access-tff6t\") pod \"nova-cell0-cell-mapping-mzcxs\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.953506 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.955965 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.958853 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.961653 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.963745 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f80e00-ac5e-4675-b349-c2aac3bd66ca-config-data\") pod \"nova-api-0\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " pod="openstack/nova-api-0" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.963800 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f80e00-ac5e-4675-b349-c2aac3bd66ca-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " pod="openstack/nova-api-0" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.963856 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88nvc\" (UniqueName: \"kubernetes.io/projected/81f80e00-ac5e-4675-b349-c2aac3bd66ca-kube-api-access-88nvc\") pod \"nova-api-0\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " pod="openstack/nova-api-0" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.963879 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81f80e00-ac5e-4675-b349-c2aac3bd66ca-logs\") pod \"nova-api-0\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " pod="openstack/nova-api-0" Oct 02 10:03:27 crc kubenswrapper[4771]: I1002 10:03:27.997779 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.077175 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.078352 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f80e00-ac5e-4675-b349-c2aac3bd66ca-config-data\") pod \"nova-api-0\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " pod="openstack/nova-api-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.078580 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f80e00-ac5e-4675-b349-c2aac3bd66ca-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " pod="openstack/nova-api-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.078707 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v76hk\" (UniqueName: \"kubernetes.io/projected/8660c7e8-08fa-49fc-ab03-f0346caedce4-kube-api-access-v76hk\") pod \"nova-metadata-0\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " pod="openstack/nova-metadata-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.078854 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88nvc\" (UniqueName: \"kubernetes.io/projected/81f80e00-ac5e-4675-b349-c2aac3bd66ca-kube-api-access-88nvc\") pod \"nova-api-0\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " pod="openstack/nova-api-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.078904 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81f80e00-ac5e-4675-b349-c2aac3bd66ca-logs\") pod \"nova-api-0\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " pod="openstack/nova-api-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.079001 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msmrn\" (UniqueName: \"kubernetes.io/projected/ff529351-3126-40ed-9d60-87959d520219-kube-api-access-msmrn\") pod \"nova-scheduler-0\" (UID: \"ff529351-3126-40ed-9d60-87959d520219\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.085293 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-config-data\") pod \"nova-scheduler-0\" (UID: \"ff529351-3126-40ed-9d60-87959d520219\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.085525 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8660c7e8-08fa-49fc-ab03-f0346caedce4-config-data\") pod \"nova-metadata-0\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " pod="openstack/nova-metadata-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.085583 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8660c7e8-08fa-49fc-ab03-f0346caedce4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " pod="openstack/nova-metadata-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.085660 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ff529351-3126-40ed-9d60-87959d520219\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.085724 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8660c7e8-08fa-49fc-ab03-f0346caedce4-logs\") pod \"nova-metadata-0\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " pod="openstack/nova-metadata-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.104304 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f80e00-ac5e-4675-b349-c2aac3bd66ca-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " pod="openstack/nova-api-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.105779 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81f80e00-ac5e-4675-b349-c2aac3bd66ca-logs\") pod \"nova-api-0\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " pod="openstack/nova-api-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.114262 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f80e00-ac5e-4675-b349-c2aac3bd66ca-config-data\") pod \"nova-api-0\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " pod="openstack/nova-api-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.190591 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88nvc\" (UniqueName: \"kubernetes.io/projected/81f80e00-ac5e-4675-b349-c2aac3bd66ca-kube-api-access-88nvc\") pod \"nova-api-0\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " pod="openstack/nova-api-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.192028 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msmrn\" (UniqueName: \"kubernetes.io/projected/ff529351-3126-40ed-9d60-87959d520219-kube-api-access-msmrn\") pod \"nova-scheduler-0\" (UID: \"ff529351-3126-40ed-9d60-87959d520219\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.192192 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-config-data\") pod \"nova-scheduler-0\" (UID: \"ff529351-3126-40ed-9d60-87959d520219\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.192335 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8660c7e8-08fa-49fc-ab03-f0346caedce4-config-data\") pod \"nova-metadata-0\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " pod="openstack/nova-metadata-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.192390 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8660c7e8-08fa-49fc-ab03-f0346caedce4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " pod="openstack/nova-metadata-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.192453 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ff529351-3126-40ed-9d60-87959d520219\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.192503 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8660c7e8-08fa-49fc-ab03-f0346caedce4-logs\") pod \"nova-metadata-0\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " pod="openstack/nova-metadata-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.192795 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v76hk\" (UniqueName: \"kubernetes.io/projected/8660c7e8-08fa-49fc-ab03-f0346caedce4-kube-api-access-v76hk\") pod \"nova-metadata-0\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " pod="openstack/nova-metadata-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.195499 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8660c7e8-08fa-49fc-ab03-f0346caedce4-logs\") pod \"nova-metadata-0\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " pod="openstack/nova-metadata-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.201395 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8660c7e8-08fa-49fc-ab03-f0346caedce4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " pod="openstack/nova-metadata-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.203342 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-config-data\") pod \"nova-scheduler-0\" (UID: \"ff529351-3126-40ed-9d60-87959d520219\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.209624 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ff529351-3126-40ed-9d60-87959d520219\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.231389 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8660c7e8-08fa-49fc-ab03-f0346caedce4-config-data\") pod \"nova-metadata-0\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " pod="openstack/nova-metadata-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.232090 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msmrn\" (UniqueName: \"kubernetes.io/projected/ff529351-3126-40ed-9d60-87959d520219-kube-api-access-msmrn\") pod \"nova-scheduler-0\" (UID: \"ff529351-3126-40ed-9d60-87959d520219\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.242089 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v76hk\" (UniqueName: \"kubernetes.io/projected/8660c7e8-08fa-49fc-ab03-f0346caedce4-kube-api-access-v76hk\") pod \"nova-metadata-0\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " pod="openstack/nova-metadata-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.309959 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.312317 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.317607 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.326256 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-gphqv"] Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.331751 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.341701 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.342864 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.378363 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.378979 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.393061 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-gphqv"] Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.408402 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab155770-4f5d-43cf-a234-7c9394177ef6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab155770-4f5d-43cf-a234-7c9394177ef6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.408500 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.408546 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4zph\" (UniqueName: \"kubernetes.io/projected/7ccdd3d7-717c-4879-9155-41669900240b-kube-api-access-f4zph\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.408605 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab155770-4f5d-43cf-a234-7c9394177ef6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab155770-4f5d-43cf-a234-7c9394177ef6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.408643 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.408673 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrg9h\" (UniqueName: \"kubernetes.io/projected/ab155770-4f5d-43cf-a234-7c9394177ef6-kube-api-access-vrg9h\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab155770-4f5d-43cf-a234-7c9394177ef6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.408728 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-dns-svc\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.418866 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-config\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.439011 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.470626 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-afdd-account-create-5znqf"] Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.473008 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-afdd-account-create-5znqf" Oct 02 10:03:28 crc kubenswrapper[4771]: I1002 10:03:28.480538 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.487315 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-afdd-account-create-5znqf"] Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.543721 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab155770-4f5d-43cf-a234-7c9394177ef6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab155770-4f5d-43cf-a234-7c9394177ef6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.544000 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.544030 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4zph\" (UniqueName: \"kubernetes.io/projected/7ccdd3d7-717c-4879-9155-41669900240b-kube-api-access-f4zph\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.544064 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab155770-4f5d-43cf-a234-7c9394177ef6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab155770-4f5d-43cf-a234-7c9394177ef6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.544091 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.544105 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrg9h\" (UniqueName: \"kubernetes.io/projected/ab155770-4f5d-43cf-a234-7c9394177ef6-kube-api-access-vrg9h\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab155770-4f5d-43cf-a234-7c9394177ef6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.544226 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-dns-svc\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.544299 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-config\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.544335 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.551450 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.557517 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.561358 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-dns-svc\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.562539 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.566121 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab155770-4f5d-43cf-a234-7c9394177ef6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab155770-4f5d-43cf-a234-7c9394177ef6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.568516 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-config\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.571383 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab155770-4f5d-43cf-a234-7c9394177ef6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab155770-4f5d-43cf-a234-7c9394177ef6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.579938 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrg9h\" (UniqueName: \"kubernetes.io/projected/ab155770-4f5d-43cf-a234-7c9394177ef6-kube-api-access-vrg9h\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab155770-4f5d-43cf-a234-7c9394177ef6\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.591019 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4zph\" (UniqueName: \"kubernetes.io/projected/7ccdd3d7-717c-4879-9155-41669900240b-kube-api-access-f4zph\") pod \"dnsmasq-dns-9b86998b5-gphqv\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.653486 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t88xg\" (UniqueName: \"kubernetes.io/projected/01978d22-b875-4359-8447-e96168822a2a-kube-api-access-t88xg\") pod \"aodh-afdd-account-create-5znqf\" (UID: \"01978d22-b875-4359-8447-e96168822a2a\") " pod="openstack/aodh-afdd-account-create-5znqf" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.667872 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.733469 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.756460 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t88xg\" (UniqueName: \"kubernetes.io/projected/01978d22-b875-4359-8447-e96168822a2a-kube-api-access-t88xg\") pod \"aodh-afdd-account-create-5znqf\" (UID: \"01978d22-b875-4359-8447-e96168822a2a\") " pod="openstack/aodh-afdd-account-create-5znqf" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.785337 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t88xg\" (UniqueName: \"kubernetes.io/projected/01978d22-b875-4359-8447-e96168822a2a-kube-api-access-t88xg\") pod \"aodh-afdd-account-create-5znqf\" (UID: \"01978d22-b875-4359-8447-e96168822a2a\") " pod="openstack/aodh-afdd-account-create-5znqf" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.833529 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-afdd-account-create-5znqf" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:28.932799 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-mzcxs"] Oct 02 10:03:29 crc kubenswrapper[4771]: W1002 10:03:29.040279 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod913531c6_8fa3_4c33_87a4_22f18bb251b8.slice/crio-47edced14445e05cf4580848ebbc81d7caf922d3818f1ad2d621bb05ffad5017 WatchSource:0}: Error finding container 47edced14445e05cf4580848ebbc81d7caf922d3818f1ad2d621bb05ffad5017: Status 404 returned error can't find the container with id 47edced14445e05cf4580848ebbc81d7caf922d3818f1ad2d621bb05ffad5017 Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.129661 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-djw9k"] Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.145839 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.151733 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.151923 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.153719 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-djw9k"] Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.275289 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-djw9k\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.275541 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfjfd\" (UniqueName: \"kubernetes.io/projected/729318ed-7618-49be-bcd1-43734b293975-kube-api-access-xfjfd\") pod \"nova-cell1-conductor-db-sync-djw9k\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.275670 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-config-data\") pod \"nova-cell1-conductor-db-sync-djw9k\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.275784 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-scripts\") pod \"nova-cell1-conductor-db-sync-djw9k\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.378223 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-scripts\") pod \"nova-cell1-conductor-db-sync-djw9k\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.378731 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-djw9k\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.378942 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfjfd\" (UniqueName: \"kubernetes.io/projected/729318ed-7618-49be-bcd1-43734b293975-kube-api-access-xfjfd\") pod \"nova-cell1-conductor-db-sync-djw9k\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.378971 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-config-data\") pod \"nova-cell1-conductor-db-sync-djw9k\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.385550 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-djw9k\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.391782 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-scripts\") pod \"nova-cell1-conductor-db-sync-djw9k\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.392484 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-config-data\") pod \"nova-cell1-conductor-db-sync-djw9k\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.400005 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfjfd\" (UniqueName: \"kubernetes.io/projected/729318ed-7618-49be-bcd1-43734b293975-kube-api-access-xfjfd\") pod \"nova-cell1-conductor-db-sync-djw9k\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.542981 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.871920 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mzcxs" event={"ID":"913531c6-8fa3-4c33-87a4-22f18bb251b8","Type":"ContainerStarted","Data":"0ccc59478bfea768a2f882afbb3e5f6d0260e6d4a3d96c01bb215769d830ea0a"} Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.872236 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mzcxs" event={"ID":"913531c6-8fa3-4c33-87a4-22f18bb251b8","Type":"ContainerStarted","Data":"47edced14445e05cf4580848ebbc81d7caf922d3818f1ad2d621bb05ffad5017"} Oct 02 10:03:29 crc kubenswrapper[4771]: I1002 10:03:29.908455 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-mzcxs" podStartSLOduration=2.908427364 podStartE2EDuration="2.908427364s" podCreationTimestamp="2025-10-02 10:03:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:03:29.893336285 +0000 UTC m=+1597.541021352" watchObservedRunningTime="2025-10-02 10:03:29.908427364 +0000 UTC m=+1597.556112431" Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.505849 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.519385 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:03:30 crc kubenswrapper[4771]: W1002 10:03:30.534070 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab155770_4f5d_43cf_a234_7c9394177ef6.slice/crio-b199e517f6494b7058f19acf4963902dc80ffe473451cfe0c5b704e07c42c3bf WatchSource:0}: Error finding container b199e517f6494b7058f19acf4963902dc80ffe473451cfe0c5b704e07c42c3bf: Status 404 returned error can't find the container with id b199e517f6494b7058f19acf4963902dc80ffe473451cfe0c5b704e07c42c3bf Oct 02 10:03:30 crc kubenswrapper[4771]: W1002 10:03:30.586428 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81f80e00_ac5e_4675_b349_c2aac3bd66ca.slice/crio-a5821106939a2e2257078ed0fd1aa46316b837faac90ef7069e5a97d581219d7 WatchSource:0}: Error finding container a5821106939a2e2257078ed0fd1aa46316b837faac90ef7069e5a97d581219d7: Status 404 returned error can't find the container with id a5821106939a2e2257078ed0fd1aa46316b837faac90ef7069e5a97d581219d7 Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.598202 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.626106 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-afdd-account-create-5znqf"] Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.662584 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.717198 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-gphqv"] Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.769944 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-djw9k"] Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.934048 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-djw9k" event={"ID":"729318ed-7618-49be-bcd1-43734b293975","Type":"ContainerStarted","Data":"dd5930a947e0cc04098f3a2a51f8d61702f3e2e3e441dee1127a29b809308604"} Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.936828 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-afdd-account-create-5znqf" event={"ID":"01978d22-b875-4359-8447-e96168822a2a","Type":"ContainerStarted","Data":"8251d7f540d58d6ea25e5e6a403a908c8de76e25898795d2b131a24440786ba5"} Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.941203 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ff529351-3126-40ed-9d60-87959d520219","Type":"ContainerStarted","Data":"abcc81ff8aac70a133b6b5adb36daf0b68556f0e59a78f41ba913285f95ceee0"} Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.949770 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ab155770-4f5d-43cf-a234-7c9394177ef6","Type":"ContainerStarted","Data":"b199e517f6494b7058f19acf4963902dc80ffe473451cfe0c5b704e07c42c3bf"} Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.960516 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" event={"ID":"7ccdd3d7-717c-4879-9155-41669900240b","Type":"ContainerStarted","Data":"8be4f00e45f1892f0d13272e956df3718b990d43e014fa8f5cdf6023b849dd54"} Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.969869 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8660c7e8-08fa-49fc-ab03-f0346caedce4","Type":"ContainerStarted","Data":"c420ac02ef9d8cc7bda54008ec6740a952e4643409b0152bcfd2db0d296ab618"} Oct 02 10:03:30 crc kubenswrapper[4771]: I1002 10:03:30.988144 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f80e00-ac5e-4675-b349-c2aac3bd66ca","Type":"ContainerStarted","Data":"a5821106939a2e2257078ed0fd1aa46316b837faac90ef7069e5a97d581219d7"} Oct 02 10:03:32 crc kubenswrapper[4771]: I1002 10:03:32.013658 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-djw9k" event={"ID":"729318ed-7618-49be-bcd1-43734b293975","Type":"ContainerStarted","Data":"dc73f360304fbd06b30fb404084e0f2ef2aae0cdca71095d21d610bfb16c249e"} Oct 02 10:03:32 crc kubenswrapper[4771]: I1002 10:03:32.029541 4771 generic.go:334] "Generic (PLEG): container finished" podID="01978d22-b875-4359-8447-e96168822a2a" containerID="9165c06caa4386e66c7ead54f76dac3f12ac81c3c8c98c541908d959ccb9ea48" exitCode=0 Oct 02 10:03:32 crc kubenswrapper[4771]: I1002 10:03:32.029612 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-afdd-account-create-5znqf" event={"ID":"01978d22-b875-4359-8447-e96168822a2a","Type":"ContainerDied","Data":"9165c06caa4386e66c7ead54f76dac3f12ac81c3c8c98c541908d959ccb9ea48"} Oct 02 10:03:32 crc kubenswrapper[4771]: I1002 10:03:32.037971 4771 generic.go:334] "Generic (PLEG): container finished" podID="7ccdd3d7-717c-4879-9155-41669900240b" containerID="b387454540528deec2342f4d808d6306bdb8825a835d5715c0efe51766e1b363" exitCode=0 Oct 02 10:03:32 crc kubenswrapper[4771]: I1002 10:03:32.038040 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" event={"ID":"7ccdd3d7-717c-4879-9155-41669900240b","Type":"ContainerDied","Data":"b387454540528deec2342f4d808d6306bdb8825a835d5715c0efe51766e1b363"} Oct 02 10:03:32 crc kubenswrapper[4771]: I1002 10:03:32.047430 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-djw9k" podStartSLOduration=3.047398588 podStartE2EDuration="3.047398588s" podCreationTimestamp="2025-10-02 10:03:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:03:32.032665979 +0000 UTC m=+1599.680351056" watchObservedRunningTime="2025-10-02 10:03:32.047398588 +0000 UTC m=+1599.695083655" Oct 02 10:03:32 crc kubenswrapper[4771]: I1002 10:03:32.681190 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:03:32 crc kubenswrapper[4771]: I1002 10:03:32.693571 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:03:33 crc kubenswrapper[4771]: I1002 10:03:33.060527 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" event={"ID":"7ccdd3d7-717c-4879-9155-41669900240b","Type":"ContainerStarted","Data":"884149544bd01e9f83b4a17b66cd975a805ee6d96fc50f266ff9351ca1217d87"} Oct 02 10:03:33 crc kubenswrapper[4771]: I1002 10:03:33.060914 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:33 crc kubenswrapper[4771]: I1002 10:03:33.088296 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" podStartSLOduration=5.088276931 podStartE2EDuration="5.088276931s" podCreationTimestamp="2025-10-02 10:03:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:03:33.083467515 +0000 UTC m=+1600.731152592" watchObservedRunningTime="2025-10-02 10:03:33.088276931 +0000 UTC m=+1600.735961998" Oct 02 10:03:34 crc kubenswrapper[4771]: I1002 10:03:34.667818 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-afdd-account-create-5znqf" Oct 02 10:03:34 crc kubenswrapper[4771]: I1002 10:03:34.815458 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t88xg\" (UniqueName: \"kubernetes.io/projected/01978d22-b875-4359-8447-e96168822a2a-kube-api-access-t88xg\") pod \"01978d22-b875-4359-8447-e96168822a2a\" (UID: \"01978d22-b875-4359-8447-e96168822a2a\") " Oct 02 10:03:34 crc kubenswrapper[4771]: I1002 10:03:34.823699 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01978d22-b875-4359-8447-e96168822a2a-kube-api-access-t88xg" (OuterVolumeSpecName: "kube-api-access-t88xg") pod "01978d22-b875-4359-8447-e96168822a2a" (UID: "01978d22-b875-4359-8447-e96168822a2a"). InnerVolumeSpecName "kube-api-access-t88xg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:34 crc kubenswrapper[4771]: I1002 10:03:34.919675 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t88xg\" (UniqueName: \"kubernetes.io/projected/01978d22-b875-4359-8447-e96168822a2a-kube-api-access-t88xg\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:35 crc kubenswrapper[4771]: I1002 10:03:35.100875 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-afdd-account-create-5znqf" event={"ID":"01978d22-b875-4359-8447-e96168822a2a","Type":"ContainerDied","Data":"8251d7f540d58d6ea25e5e6a403a908c8de76e25898795d2b131a24440786ba5"} Oct 02 10:03:35 crc kubenswrapper[4771]: I1002 10:03:35.100934 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8251d7f540d58d6ea25e5e6a403a908c8de76e25898795d2b131a24440786ba5" Oct 02 10:03:35 crc kubenswrapper[4771]: I1002 10:03:35.101004 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-afdd-account-create-5znqf" Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.114669 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ab155770-4f5d-43cf-a234-7c9394177ef6","Type":"ContainerStarted","Data":"28d94e68497f8ab6a78649fc80178b2cdf9e7795198e38f81dfff5002f70f06a"} Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.114731 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="ab155770-4f5d-43cf-a234-7c9394177ef6" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://28d94e68497f8ab6a78649fc80178b2cdf9e7795198e38f81dfff5002f70f06a" gracePeriod=30 Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.116809 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8660c7e8-08fa-49fc-ab03-f0346caedce4","Type":"ContainerStarted","Data":"e96030dfcba88964dbcab13df91d670250e729241dd3b3f7f1f3596f3d1c2b8d"} Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.116854 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8660c7e8-08fa-49fc-ab03-f0346caedce4","Type":"ContainerStarted","Data":"ff7920741eb56baa8292368ff934ac028cb877c2d4fc8aa93fcd48c6a1af15fb"} Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.116986 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8660c7e8-08fa-49fc-ab03-f0346caedce4" containerName="nova-metadata-log" containerID="cri-o://ff7920741eb56baa8292368ff934ac028cb877c2d4fc8aa93fcd48c6a1af15fb" gracePeriod=30 Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.117096 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8660c7e8-08fa-49fc-ab03-f0346caedce4" containerName="nova-metadata-metadata" containerID="cri-o://e96030dfcba88964dbcab13df91d670250e729241dd3b3f7f1f3596f3d1c2b8d" gracePeriod=30 Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.127147 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f80e00-ac5e-4675-b349-c2aac3bd66ca","Type":"ContainerStarted","Data":"efb8a247531905802a42ad37979442fd1c353a30b51215fd89086ae81d6a8160"} Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.127198 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f80e00-ac5e-4675-b349-c2aac3bd66ca","Type":"ContainerStarted","Data":"b235c76c9634c5c3950bf7b0a33f1cd9f004970623e6d7e9a619706ae9a470ec"} Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.130810 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ff529351-3126-40ed-9d60-87959d520219","Type":"ContainerStarted","Data":"2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285"} Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.144249 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.390374312 podStartE2EDuration="8.144227315s" podCreationTimestamp="2025-10-02 10:03:28 +0000 UTC" firstStartedPulling="2025-10-02 10:03:30.542359194 +0000 UTC m=+1598.190044261" lastFinishedPulling="2025-10-02 10:03:35.296212197 +0000 UTC m=+1602.943897264" observedRunningTime="2025-10-02 10:03:36.136694281 +0000 UTC m=+1603.784379358" watchObservedRunningTime="2025-10-02 10:03:36.144227315 +0000 UTC m=+1603.791912382" Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.169600 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=4.59424478 podStartE2EDuration="9.169577626s" podCreationTimestamp="2025-10-02 10:03:27 +0000 UTC" firstStartedPulling="2025-10-02 10:03:30.542824267 +0000 UTC m=+1598.190509334" lastFinishedPulling="2025-10-02 10:03:35.118157113 +0000 UTC m=+1602.765842180" observedRunningTime="2025-10-02 10:03:36.156136304 +0000 UTC m=+1603.803821371" watchObservedRunningTime="2025-10-02 10:03:36.169577626 +0000 UTC m=+1603.817262693" Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.187044 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.49191263 podStartE2EDuration="9.187019372s" podCreationTimestamp="2025-10-02 10:03:27 +0000 UTC" firstStartedPulling="2025-10-02 10:03:30.597139402 +0000 UTC m=+1598.244824469" lastFinishedPulling="2025-10-02 10:03:35.292246134 +0000 UTC m=+1602.939931211" observedRunningTime="2025-10-02 10:03:36.176251906 +0000 UTC m=+1603.823936973" watchObservedRunningTime="2025-10-02 10:03:36.187019372 +0000 UTC m=+1603.834704439" Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.203620 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.478939702 podStartE2EDuration="9.203598224s" podCreationTimestamp="2025-10-02 10:03:27 +0000 UTC" firstStartedPulling="2025-10-02 10:03:30.627166556 +0000 UTC m=+1598.274851623" lastFinishedPulling="2025-10-02 10:03:35.351825078 +0000 UTC m=+1602.999510145" observedRunningTime="2025-10-02 10:03:36.200680971 +0000 UTC m=+1603.848366058" watchObservedRunningTime="2025-10-02 10:03:36.203598224 +0000 UTC m=+1603.851283291" Oct 02 10:03:36 crc kubenswrapper[4771]: I1002 10:03:36.681258 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:03:36 crc kubenswrapper[4771]: E1002 10:03:36.681684 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:03:37 crc kubenswrapper[4771]: I1002 10:03:37.143134 4771 generic.go:334] "Generic (PLEG): container finished" podID="8660c7e8-08fa-49fc-ab03-f0346caedce4" containerID="ff7920741eb56baa8292368ff934ac028cb877c2d4fc8aa93fcd48c6a1af15fb" exitCode=143 Oct 02 10:03:37 crc kubenswrapper[4771]: I1002 10:03:37.143176 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8660c7e8-08fa-49fc-ab03-f0346caedce4","Type":"ContainerDied","Data":"ff7920741eb56baa8292368ff934ac028cb877c2d4fc8aa93fcd48c6a1af15fb"} Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.332667 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.333715 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.344593 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.345032 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.380016 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.380072 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.415285 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.668164 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.735321 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.809607 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-8xp24"] Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.809854 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" podUID="0ec19f19-2184-43e0-b683-d35db4a08da0" containerName="dnsmasq-dns" containerID="cri-o://4c75700c18b034995223129b8b2d3e4c4146fae13fe3a9e7558ab848cc004eff" gracePeriod=10 Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.929172 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-6mtn4"] Oct 02 10:03:38 crc kubenswrapper[4771]: E1002 10:03:38.929926 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01978d22-b875-4359-8447-e96168822a2a" containerName="mariadb-account-create" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.929948 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="01978d22-b875-4359-8447-e96168822a2a" containerName="mariadb-account-create" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.930271 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="01978d22-b875-4359-8447-e96168822a2a" containerName="mariadb-account-create" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.931476 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.937651 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.937787 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-mnlz2" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.938001 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 02 10:03:38 crc kubenswrapper[4771]: I1002 10:03:38.945326 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-6mtn4"] Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.049972 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-scripts\") pod \"aodh-db-sync-6mtn4\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.050083 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5lcq\" (UniqueName: \"kubernetes.io/projected/9d8b4155-64a1-4c45-b368-72013bcdb7aa-kube-api-access-t5lcq\") pod \"aodh-db-sync-6mtn4\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.050158 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-combined-ca-bundle\") pod \"aodh-db-sync-6mtn4\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.050210 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-config-data\") pod \"aodh-db-sync-6mtn4\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.152951 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-scripts\") pod \"aodh-db-sync-6mtn4\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.153469 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5lcq\" (UniqueName: \"kubernetes.io/projected/9d8b4155-64a1-4c45-b368-72013bcdb7aa-kube-api-access-t5lcq\") pod \"aodh-db-sync-6mtn4\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.153521 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-combined-ca-bundle\") pod \"aodh-db-sync-6mtn4\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.153598 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-config-data\") pod \"aodh-db-sync-6mtn4\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.166694 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-combined-ca-bundle\") pod \"aodh-db-sync-6mtn4\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.167191 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-config-data\") pod \"aodh-db-sync-6mtn4\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.170595 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-scripts\") pod \"aodh-db-sync-6mtn4\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.180781 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5lcq\" (UniqueName: \"kubernetes.io/projected/9d8b4155-64a1-4c45-b368-72013bcdb7aa-kube-api-access-t5lcq\") pod \"aodh-db-sync-6mtn4\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.208413 4771 generic.go:334] "Generic (PLEG): container finished" podID="0ec19f19-2184-43e0-b683-d35db4a08da0" containerID="4c75700c18b034995223129b8b2d3e4c4146fae13fe3a9e7558ab848cc004eff" exitCode=0 Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.208519 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" event={"ID":"0ec19f19-2184-43e0-b683-d35db4a08da0","Type":"ContainerDied","Data":"4c75700c18b034995223129b8b2d3e4c4146fae13fe3a9e7558ab848cc004eff"} Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.270910 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.352095 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.457381 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="81f80e00-ac5e-4675-b349-c2aac3bd66ca" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.238:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.458117 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="81f80e00-ac5e-4675-b349-c2aac3bd66ca" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.238:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.716910 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.809165 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-ovsdbserver-nb\") pod \"0ec19f19-2184-43e0-b683-d35db4a08da0\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.809260 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-dns-svc\") pod \"0ec19f19-2184-43e0-b683-d35db4a08da0\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.809314 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-dns-swift-storage-0\") pod \"0ec19f19-2184-43e0-b683-d35db4a08da0\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.809367 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l94dk\" (UniqueName: \"kubernetes.io/projected/0ec19f19-2184-43e0-b683-d35db4a08da0-kube-api-access-l94dk\") pod \"0ec19f19-2184-43e0-b683-d35db4a08da0\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.809396 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-ovsdbserver-sb\") pod \"0ec19f19-2184-43e0-b683-d35db4a08da0\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.809486 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-config\") pod \"0ec19f19-2184-43e0-b683-d35db4a08da0\" (UID: \"0ec19f19-2184-43e0-b683-d35db4a08da0\") " Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.818201 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ec19f19-2184-43e0-b683-d35db4a08da0-kube-api-access-l94dk" (OuterVolumeSpecName: "kube-api-access-l94dk") pod "0ec19f19-2184-43e0-b683-d35db4a08da0" (UID: "0ec19f19-2184-43e0-b683-d35db4a08da0"). InnerVolumeSpecName "kube-api-access-l94dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.916253 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l94dk\" (UniqueName: \"kubernetes.io/projected/0ec19f19-2184-43e0-b683-d35db4a08da0-kube-api-access-l94dk\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.917001 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0ec19f19-2184-43e0-b683-d35db4a08da0" (UID: "0ec19f19-2184-43e0-b683-d35db4a08da0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.941675 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0ec19f19-2184-43e0-b683-d35db4a08da0" (UID: "0ec19f19-2184-43e0-b683-d35db4a08da0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.952626 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0ec19f19-2184-43e0-b683-d35db4a08da0" (UID: "0ec19f19-2184-43e0-b683-d35db4a08da0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:03:39 crc kubenswrapper[4771]: I1002 10:03:39.990233 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0ec19f19-2184-43e0-b683-d35db4a08da0" (UID: "0ec19f19-2184-43e0-b683-d35db4a08da0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.022255 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.022306 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.022320 4771 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.022333 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.035816 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-config" (OuterVolumeSpecName: "config") pod "0ec19f19-2184-43e0-b683-d35db4a08da0" (UID: "0ec19f19-2184-43e0-b683-d35db4a08da0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.046642 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-6mtn4"] Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.126511 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ec19f19-2184-43e0-b683-d35db4a08da0-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.233258 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6mtn4" event={"ID":"9d8b4155-64a1-4c45-b368-72013bcdb7aa","Type":"ContainerStarted","Data":"7f01ffdf26e237f3d0fd0398f215f00865cbcf8543b8aed5ca140a85f1a41f9a"} Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.244897 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.258627 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-8xp24" event={"ID":"0ec19f19-2184-43e0-b683-d35db4a08da0","Type":"ContainerDied","Data":"f987c0fe29646178d6c930a015339d9e525eba607468582c0b206d53937fe2fe"} Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.258730 4771 scope.go:117] "RemoveContainer" containerID="4c75700c18b034995223129b8b2d3e4c4146fae13fe3a9e7558ab848cc004eff" Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.331225 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-8xp24"] Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.337258 4771 scope.go:117] "RemoveContainer" containerID="ef2ceb84f86437104a92e490ba4ddae48fcb8417a02b16deac0e333d9b6aaaa7" Oct 02 10:03:40 crc kubenswrapper[4771]: I1002 10:03:40.344234 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-8xp24"] Oct 02 10:03:40 crc kubenswrapper[4771]: E1002 10:03:40.615414 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ec19f19_2184_43e0_b683_d35db4a08da0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod913531c6_8fa3_4c33_87a4_22f18bb251b8.slice/crio-0ccc59478bfea768a2f882afbb3e5f6d0260e6d4a3d96c01bb215769d830ea0a.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:03:41 crc kubenswrapper[4771]: I1002 10:03:41.261579 4771 generic.go:334] "Generic (PLEG): container finished" podID="913531c6-8fa3-4c33-87a4-22f18bb251b8" containerID="0ccc59478bfea768a2f882afbb3e5f6d0260e6d4a3d96c01bb215769d830ea0a" exitCode=0 Oct 02 10:03:41 crc kubenswrapper[4771]: I1002 10:03:41.261677 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mzcxs" event={"ID":"913531c6-8fa3-4c33-87a4-22f18bb251b8","Type":"ContainerDied","Data":"0ccc59478bfea768a2f882afbb3e5f6d0260e6d4a3d96c01bb215769d830ea0a"} Oct 02 10:03:41 crc kubenswrapper[4771]: I1002 10:03:41.271263 4771 generic.go:334] "Generic (PLEG): container finished" podID="729318ed-7618-49be-bcd1-43734b293975" containerID="dc73f360304fbd06b30fb404084e0f2ef2aae0cdca71095d21d610bfb16c249e" exitCode=0 Oct 02 10:03:41 crc kubenswrapper[4771]: I1002 10:03:41.271582 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-djw9k" event={"ID":"729318ed-7618-49be-bcd1-43734b293975","Type":"ContainerDied","Data":"dc73f360304fbd06b30fb404084e0f2ef2aae0cdca71095d21d610bfb16c249e"} Oct 02 10:03:41 crc kubenswrapper[4771]: I1002 10:03:41.700995 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ec19f19-2184-43e0-b683-d35db4a08da0" path="/var/lib/kubelet/pods/0ec19f19-2184-43e0-b683-d35db4a08da0/volumes" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.093007 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.399220 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-djw9k" event={"ID":"729318ed-7618-49be-bcd1-43734b293975","Type":"ContainerDied","Data":"dd5930a947e0cc04098f3a2a51f8d61702f3e2e3e441dee1127a29b809308604"} Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.399588 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd5930a947e0cc04098f3a2a51f8d61702f3e2e3e441dee1127a29b809308604" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.401349 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mzcxs" event={"ID":"913531c6-8fa3-4c33-87a4-22f18bb251b8","Type":"ContainerDied","Data":"47edced14445e05cf4580848ebbc81d7caf922d3818f1ad2d621bb05ffad5017"} Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.401374 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47edced14445e05cf4580848ebbc81d7caf922d3818f1ad2d621bb05ffad5017" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.446569 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.459807 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.583935 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-scripts\") pod \"913531c6-8fa3-4c33-87a4-22f18bb251b8\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.583993 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-scripts\") pod \"729318ed-7618-49be-bcd1-43734b293975\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.584114 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-combined-ca-bundle\") pod \"729318ed-7618-49be-bcd1-43734b293975\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.584368 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-config-data\") pod \"913531c6-8fa3-4c33-87a4-22f18bb251b8\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.584437 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tff6t\" (UniqueName: \"kubernetes.io/projected/913531c6-8fa3-4c33-87a4-22f18bb251b8-kube-api-access-tff6t\") pod \"913531c6-8fa3-4c33-87a4-22f18bb251b8\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.584483 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-config-data\") pod \"729318ed-7618-49be-bcd1-43734b293975\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.584547 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfjfd\" (UniqueName: \"kubernetes.io/projected/729318ed-7618-49be-bcd1-43734b293975-kube-api-access-xfjfd\") pod \"729318ed-7618-49be-bcd1-43734b293975\" (UID: \"729318ed-7618-49be-bcd1-43734b293975\") " Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.584579 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-combined-ca-bundle\") pod \"913531c6-8fa3-4c33-87a4-22f18bb251b8\" (UID: \"913531c6-8fa3-4c33-87a4-22f18bb251b8\") " Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.591480 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-scripts" (OuterVolumeSpecName: "scripts") pod "729318ed-7618-49be-bcd1-43734b293975" (UID: "729318ed-7618-49be-bcd1-43734b293975"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.591722 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/913531c6-8fa3-4c33-87a4-22f18bb251b8-kube-api-access-tff6t" (OuterVolumeSpecName: "kube-api-access-tff6t") pod "913531c6-8fa3-4c33-87a4-22f18bb251b8" (UID: "913531c6-8fa3-4c33-87a4-22f18bb251b8"). InnerVolumeSpecName "kube-api-access-tff6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.592370 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/729318ed-7618-49be-bcd1-43734b293975-kube-api-access-xfjfd" (OuterVolumeSpecName: "kube-api-access-xfjfd") pod "729318ed-7618-49be-bcd1-43734b293975" (UID: "729318ed-7618-49be-bcd1-43734b293975"). InnerVolumeSpecName "kube-api-access-xfjfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.609146 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-scripts" (OuterVolumeSpecName: "scripts") pod "913531c6-8fa3-4c33-87a4-22f18bb251b8" (UID: "913531c6-8fa3-4c33-87a4-22f18bb251b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.625283 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-config-data" (OuterVolumeSpecName: "config-data") pod "729318ed-7618-49be-bcd1-43734b293975" (UID: "729318ed-7618-49be-bcd1-43734b293975"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.628946 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-config-data" (OuterVolumeSpecName: "config-data") pod "913531c6-8fa3-4c33-87a4-22f18bb251b8" (UID: "913531c6-8fa3-4c33-87a4-22f18bb251b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.629068 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "729318ed-7618-49be-bcd1-43734b293975" (UID: "729318ed-7618-49be-bcd1-43734b293975"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.634767 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "913531c6-8fa3-4c33-87a4-22f18bb251b8" (UID: "913531c6-8fa3-4c33-87a4-22f18bb251b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.687788 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.687839 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tff6t\" (UniqueName: \"kubernetes.io/projected/913531c6-8fa3-4c33-87a4-22f18bb251b8-kube-api-access-tff6t\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.687851 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.687869 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfjfd\" (UniqueName: \"kubernetes.io/projected/729318ed-7618-49be-bcd1-43734b293975-kube-api-access-xfjfd\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.687879 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.687887 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913531c6-8fa3-4c33-87a4-22f18bb251b8-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.687896 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:45 crc kubenswrapper[4771]: I1002 10:03:45.687905 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/729318ed-7618-49be-bcd1-43734b293975-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.413704 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6mtn4" event={"ID":"9d8b4155-64a1-4c45-b368-72013bcdb7aa","Type":"ContainerStarted","Data":"02a52d8f511b6949ce8ca03694d385edfbd592a3aa00254acc681dca52c62ec2"} Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.413728 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-djw9k" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.413803 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mzcxs" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.432324 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-6mtn4" podStartSLOduration=3.257958015 podStartE2EDuration="8.432306747s" podCreationTimestamp="2025-10-02 10:03:38 +0000 UTC" firstStartedPulling="2025-10-02 10:03:40.064489511 +0000 UTC m=+1607.712174578" lastFinishedPulling="2025-10-02 10:03:45.238838243 +0000 UTC m=+1612.886523310" observedRunningTime="2025-10-02 10:03:46.42993626 +0000 UTC m=+1614.077621327" watchObservedRunningTime="2025-10-02 10:03:46.432306747 +0000 UTC m=+1614.079991814" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.581057 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 10:03:46 crc kubenswrapper[4771]: E1002 10:03:46.581896 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913531c6-8fa3-4c33-87a4-22f18bb251b8" containerName="nova-manage" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.581922 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="913531c6-8fa3-4c33-87a4-22f18bb251b8" containerName="nova-manage" Oct 02 10:03:46 crc kubenswrapper[4771]: E1002 10:03:46.581944 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ec19f19-2184-43e0-b683-d35db4a08da0" containerName="init" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.581953 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ec19f19-2184-43e0-b683-d35db4a08da0" containerName="init" Oct 02 10:03:46 crc kubenswrapper[4771]: E1002 10:03:46.581966 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ec19f19-2184-43e0-b683-d35db4a08da0" containerName="dnsmasq-dns" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.581972 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ec19f19-2184-43e0-b683-d35db4a08da0" containerName="dnsmasq-dns" Oct 02 10:03:46 crc kubenswrapper[4771]: E1002 10:03:46.581985 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="729318ed-7618-49be-bcd1-43734b293975" containerName="nova-cell1-conductor-db-sync" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.581991 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="729318ed-7618-49be-bcd1-43734b293975" containerName="nova-cell1-conductor-db-sync" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.582355 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="729318ed-7618-49be-bcd1-43734b293975" containerName="nova-cell1-conductor-db-sync" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.582382 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="913531c6-8fa3-4c33-87a4-22f18bb251b8" containerName="nova-manage" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.582397 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ec19f19-2184-43e0-b683-d35db4a08da0" containerName="dnsmasq-dns" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.583625 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.587545 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.620381 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.716245 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.716568 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ff529351-3126-40ed-9d60-87959d520219" containerName="nova-scheduler-scheduler" containerID="cri-o://2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285" gracePeriod=30 Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.730918 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.731356 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="81f80e00-ac5e-4675-b349-c2aac3bd66ca" containerName="nova-api-log" containerID="cri-o://b235c76c9634c5c3950bf7b0a33f1cd9f004970623e6d7e9a619706ae9a470ec" gracePeriod=30 Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.731480 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="81f80e00-ac5e-4675-b349-c2aac3bd66ca" containerName="nova-api-api" containerID="cri-o://efb8a247531905802a42ad37979442fd1c353a30b51215fd89086ae81d6a8160" gracePeriod=30 Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.738590 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef7b8e9e-ce03-4409-986d-fa097341328c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ef7b8e9e-ce03-4409-986d-fa097341328c\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.738711 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef7b8e9e-ce03-4409-986d-fa097341328c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ef7b8e9e-ce03-4409-986d-fa097341328c\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.739005 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs6gh\" (UniqueName: \"kubernetes.io/projected/ef7b8e9e-ce03-4409-986d-fa097341328c-kube-api-access-xs6gh\") pod \"nova-cell1-conductor-0\" (UID: \"ef7b8e9e-ce03-4409-986d-fa097341328c\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.843184 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef7b8e9e-ce03-4409-986d-fa097341328c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ef7b8e9e-ce03-4409-986d-fa097341328c\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.844009 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef7b8e9e-ce03-4409-986d-fa097341328c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ef7b8e9e-ce03-4409-986d-fa097341328c\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.844229 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs6gh\" (UniqueName: \"kubernetes.io/projected/ef7b8e9e-ce03-4409-986d-fa097341328c-kube-api-access-xs6gh\") pod \"nova-cell1-conductor-0\" (UID: \"ef7b8e9e-ce03-4409-986d-fa097341328c\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.850982 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef7b8e9e-ce03-4409-986d-fa097341328c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ef7b8e9e-ce03-4409-986d-fa097341328c\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.851114 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef7b8e9e-ce03-4409-986d-fa097341328c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ef7b8e9e-ce03-4409-986d-fa097341328c\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.874572 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs6gh\" (UniqueName: \"kubernetes.io/projected/ef7b8e9e-ce03-4409-986d-fa097341328c-kube-api-access-xs6gh\") pod \"nova-cell1-conductor-0\" (UID: \"ef7b8e9e-ce03-4409-986d-fa097341328c\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:03:46 crc kubenswrapper[4771]: I1002 10:03:46.905449 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 10:03:47 crc kubenswrapper[4771]: I1002 10:03:47.429367 4771 generic.go:334] "Generic (PLEG): container finished" podID="81f80e00-ac5e-4675-b349-c2aac3bd66ca" containerID="b235c76c9634c5c3950bf7b0a33f1cd9f004970623e6d7e9a619706ae9a470ec" exitCode=143 Oct 02 10:03:47 crc kubenswrapper[4771]: I1002 10:03:47.429435 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f80e00-ac5e-4675-b349-c2aac3bd66ca","Type":"ContainerDied","Data":"b235c76c9634c5c3950bf7b0a33f1cd9f004970623e6d7e9a619706ae9a470ec"} Oct 02 10:03:47 crc kubenswrapper[4771]: I1002 10:03:47.519963 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 10:03:47 crc kubenswrapper[4771]: W1002 10:03:47.520521 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef7b8e9e_ce03_4409_986d_fa097341328c.slice/crio-5d07353826180893ac076df7b296b11767d2a98a0cc065fe24040d589319eb7c WatchSource:0}: Error finding container 5d07353826180893ac076df7b296b11767d2a98a0cc065fe24040d589319eb7c: Status 404 returned error can't find the container with id 5d07353826180893ac076df7b296b11767d2a98a0cc065fe24040d589319eb7c Oct 02 10:03:48 crc kubenswrapper[4771]: E1002 10:03:48.383859 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:03:48 crc kubenswrapper[4771]: E1002 10:03:48.386874 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:03:48 crc kubenswrapper[4771]: E1002 10:03:48.398507 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:03:48 crc kubenswrapper[4771]: E1002 10:03:48.398628 4771 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ff529351-3126-40ed-9d60-87959d520219" containerName="nova-scheduler-scheduler" Oct 02 10:03:48 crc kubenswrapper[4771]: I1002 10:03:48.445981 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ef7b8e9e-ce03-4409-986d-fa097341328c","Type":"ContainerStarted","Data":"b695e5b7d1bcdad418117b1567609429b71f68f76986aa0f0b1a25c4cc92c06a"} Oct 02 10:03:48 crc kubenswrapper[4771]: I1002 10:03:48.446032 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ef7b8e9e-ce03-4409-986d-fa097341328c","Type":"ContainerStarted","Data":"5d07353826180893ac076df7b296b11767d2a98a0cc065fe24040d589319eb7c"} Oct 02 10:03:48 crc kubenswrapper[4771]: I1002 10:03:48.446212 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 02 10:03:48 crc kubenswrapper[4771]: I1002 10:03:48.463256 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.463230369 podStartE2EDuration="2.463230369s" podCreationTimestamp="2025-10-02 10:03:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:03:48.46152634 +0000 UTC m=+1616.109211407" watchObservedRunningTime="2025-10-02 10:03:48.463230369 +0000 UTC m=+1616.110915436" Oct 02 10:03:48 crc kubenswrapper[4771]: I1002 10:03:48.688472 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:03:48 crc kubenswrapper[4771]: E1002 10:03:48.688874 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:03:50 crc kubenswrapper[4771]: I1002 10:03:50.484568 4771 generic.go:334] "Generic (PLEG): container finished" podID="9d8b4155-64a1-4c45-b368-72013bcdb7aa" containerID="02a52d8f511b6949ce8ca03694d385edfbd592a3aa00254acc681dca52c62ec2" exitCode=0 Oct 02 10:03:50 crc kubenswrapper[4771]: I1002 10:03:50.484649 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6mtn4" event={"ID":"9d8b4155-64a1-4c45-b368-72013bcdb7aa","Type":"ContainerDied","Data":"02a52d8f511b6949ce8ca03694d385edfbd592a3aa00254acc681dca52c62ec2"} Oct 02 10:03:50 crc kubenswrapper[4771]: I1002 10:03:50.488567 4771 generic.go:334] "Generic (PLEG): container finished" podID="81f80e00-ac5e-4675-b349-c2aac3bd66ca" containerID="efb8a247531905802a42ad37979442fd1c353a30b51215fd89086ae81d6a8160" exitCode=0 Oct 02 10:03:50 crc kubenswrapper[4771]: I1002 10:03:50.488600 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f80e00-ac5e-4675-b349-c2aac3bd66ca","Type":"ContainerDied","Data":"efb8a247531905802a42ad37979442fd1c353a30b51215fd89086ae81d6a8160"} Oct 02 10:03:50 crc kubenswrapper[4771]: I1002 10:03:50.987732 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.188406 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f80e00-ac5e-4675-b349-c2aac3bd66ca-combined-ca-bundle\") pod \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.188642 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88nvc\" (UniqueName: \"kubernetes.io/projected/81f80e00-ac5e-4675-b349-c2aac3bd66ca-kube-api-access-88nvc\") pod \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.188714 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f80e00-ac5e-4675-b349-c2aac3bd66ca-config-data\") pod \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.188817 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81f80e00-ac5e-4675-b349-c2aac3bd66ca-logs\") pod \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\" (UID: \"81f80e00-ac5e-4675-b349-c2aac3bd66ca\") " Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.191620 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81f80e00-ac5e-4675-b349-c2aac3bd66ca-logs" (OuterVolumeSpecName: "logs") pod "81f80e00-ac5e-4675-b349-c2aac3bd66ca" (UID: "81f80e00-ac5e-4675-b349-c2aac3bd66ca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.199866 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81f80e00-ac5e-4675-b349-c2aac3bd66ca-kube-api-access-88nvc" (OuterVolumeSpecName: "kube-api-access-88nvc") pod "81f80e00-ac5e-4675-b349-c2aac3bd66ca" (UID: "81f80e00-ac5e-4675-b349-c2aac3bd66ca"). InnerVolumeSpecName "kube-api-access-88nvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.243869 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81f80e00-ac5e-4675-b349-c2aac3bd66ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81f80e00-ac5e-4675-b349-c2aac3bd66ca" (UID: "81f80e00-ac5e-4675-b349-c2aac3bd66ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.269347 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81f80e00-ac5e-4675-b349-c2aac3bd66ca-config-data" (OuterVolumeSpecName: "config-data") pod "81f80e00-ac5e-4675-b349-c2aac3bd66ca" (UID: "81f80e00-ac5e-4675-b349-c2aac3bd66ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.293036 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f80e00-ac5e-4675-b349-c2aac3bd66ca-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.293084 4771 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81f80e00-ac5e-4675-b349-c2aac3bd66ca-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.293096 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f80e00-ac5e-4675-b349-c2aac3bd66ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.293108 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88nvc\" (UniqueName: \"kubernetes.io/projected/81f80e00-ac5e-4675-b349-c2aac3bd66ca-kube-api-access-88nvc\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.380053 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.380405 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="d8f54e64-697d-41b5-83fd-295ce5143d04" containerName="kube-state-metrics" containerID="cri-o://d08794f587f9bdf237349764132dd957e40260917b09698e06252abe9c0e6ae9" gracePeriod=30 Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.503439 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.503727 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="e37a8032-f348-4e48-ac3a-9da319680ad2" containerName="mysqld-exporter" containerID="cri-o://c0e9c71d39ee3db61f689d3ee4d53ba4357ca3415bb3eb88f0392121dbb5f8b6" gracePeriod=30 Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.523958 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.523996 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f80e00-ac5e-4675-b349-c2aac3bd66ca","Type":"ContainerDied","Data":"a5821106939a2e2257078ed0fd1aa46316b837faac90ef7069e5a97d581219d7"} Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.524074 4771 scope.go:117] "RemoveContainer" containerID="efb8a247531905802a42ad37979442fd1c353a30b51215fd89086ae81d6a8160" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.740359 4771 scope.go:117] "RemoveContainer" containerID="b235c76c9634c5c3950bf7b0a33f1cd9f004970623e6d7e9a619706ae9a470ec" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.754081 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.800236 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.888118 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 10:03:51 crc kubenswrapper[4771]: E1002 10:03:51.889595 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f80e00-ac5e-4675-b349-c2aac3bd66ca" containerName="nova-api-log" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.889655 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f80e00-ac5e-4675-b349-c2aac3bd66ca" containerName="nova-api-log" Oct 02 10:03:51 crc kubenswrapper[4771]: E1002 10:03:51.889753 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f80e00-ac5e-4675-b349-c2aac3bd66ca" containerName="nova-api-api" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.889763 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f80e00-ac5e-4675-b349-c2aac3bd66ca" containerName="nova-api-api" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.890887 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="81f80e00-ac5e-4675-b349-c2aac3bd66ca" containerName="nova-api-log" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.890929 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="81f80e00-ac5e-4675-b349-c2aac3bd66ca" containerName="nova-api-api" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.893990 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.898230 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.947554 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.952003 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78504a04-3786-414a-94f9-e7cd45d2deee-logs\") pod \"nova-api-0\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " pod="openstack/nova-api-0" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.952165 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxxnd\" (UniqueName: \"kubernetes.io/projected/78504a04-3786-414a-94f9-e7cd45d2deee-kube-api-access-wxxnd\") pod \"nova-api-0\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " pod="openstack/nova-api-0" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.952241 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78504a04-3786-414a-94f9-e7cd45d2deee-config-data\") pod \"nova-api-0\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " pod="openstack/nova-api-0" Oct 02 10:03:51 crc kubenswrapper[4771]: I1002 10:03:51.952316 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78504a04-3786-414a-94f9-e7cd45d2deee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " pod="openstack/nova-api-0" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.027934 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.057427 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78504a04-3786-414a-94f9-e7cd45d2deee-logs\") pod \"nova-api-0\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " pod="openstack/nova-api-0" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.057570 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxxnd\" (UniqueName: \"kubernetes.io/projected/78504a04-3786-414a-94f9-e7cd45d2deee-kube-api-access-wxxnd\") pod \"nova-api-0\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " pod="openstack/nova-api-0" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.057642 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78504a04-3786-414a-94f9-e7cd45d2deee-config-data\") pod \"nova-api-0\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " pod="openstack/nova-api-0" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.057719 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78504a04-3786-414a-94f9-e7cd45d2deee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " pod="openstack/nova-api-0" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.059089 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78504a04-3786-414a-94f9-e7cd45d2deee-logs\") pod \"nova-api-0\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " pod="openstack/nova-api-0" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.066328 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78504a04-3786-414a-94f9-e7cd45d2deee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " pod="openstack/nova-api-0" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.066364 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78504a04-3786-414a-94f9-e7cd45d2deee-config-data\") pod \"nova-api-0\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " pod="openstack/nova-api-0" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.077002 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxxnd\" (UniqueName: \"kubernetes.io/projected/78504a04-3786-414a-94f9-e7cd45d2deee-kube-api-access-wxxnd\") pod \"nova-api-0\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " pod="openstack/nova-api-0" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.160795 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-scripts\") pod \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.160880 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5lcq\" (UniqueName: \"kubernetes.io/projected/9d8b4155-64a1-4c45-b368-72013bcdb7aa-kube-api-access-t5lcq\") pod \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.161043 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-config-data\") pod \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.161079 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-combined-ca-bundle\") pod \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\" (UID: \"9d8b4155-64a1-4c45-b368-72013bcdb7aa\") " Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.174499 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d8b4155-64a1-4c45-b368-72013bcdb7aa-kube-api-access-t5lcq" (OuterVolumeSpecName: "kube-api-access-t5lcq") pod "9d8b4155-64a1-4c45-b368-72013bcdb7aa" (UID: "9d8b4155-64a1-4c45-b368-72013bcdb7aa"). InnerVolumeSpecName "kube-api-access-t5lcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.174633 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-scripts" (OuterVolumeSpecName: "scripts") pod "9d8b4155-64a1-4c45-b368-72013bcdb7aa" (UID: "9d8b4155-64a1-4c45-b368-72013bcdb7aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.202455 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-config-data" (OuterVolumeSpecName: "config-data") pod "9d8b4155-64a1-4c45-b368-72013bcdb7aa" (UID: "9d8b4155-64a1-4c45-b368-72013bcdb7aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.225800 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d8b4155-64a1-4c45-b368-72013bcdb7aa" (UID: "9d8b4155-64a1-4c45-b368-72013bcdb7aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.248988 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.267204 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.267246 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.267258 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d8b4155-64a1-4c45-b368-72013bcdb7aa-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.267267 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5lcq\" (UniqueName: \"kubernetes.io/projected/9d8b4155-64a1-4c45-b368-72013bcdb7aa-kube-api-access-t5lcq\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.551188 4771 generic.go:334] "Generic (PLEG): container finished" podID="e37a8032-f348-4e48-ac3a-9da319680ad2" containerID="c0e9c71d39ee3db61f689d3ee4d53ba4357ca3415bb3eb88f0392121dbb5f8b6" exitCode=2 Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.551603 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"e37a8032-f348-4e48-ac3a-9da319680ad2","Type":"ContainerDied","Data":"c0e9c71d39ee3db61f689d3ee4d53ba4357ca3415bb3eb88f0392121dbb5f8b6"} Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.557856 4771 generic.go:334] "Generic (PLEG): container finished" podID="d8f54e64-697d-41b5-83fd-295ce5143d04" containerID="d08794f587f9bdf237349764132dd957e40260917b09698e06252abe9c0e6ae9" exitCode=2 Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.557924 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d8f54e64-697d-41b5-83fd-295ce5143d04","Type":"ContainerDied","Data":"d08794f587f9bdf237349764132dd957e40260917b09698e06252abe9c0e6ae9"} Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.569339 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6mtn4" event={"ID":"9d8b4155-64a1-4c45-b368-72013bcdb7aa","Type":"ContainerDied","Data":"7f01ffdf26e237f3d0fd0398f215f00865cbcf8543b8aed5ca140a85f1a41f9a"} Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.569392 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f01ffdf26e237f3d0fd0398f215f00865cbcf8543b8aed5ca140a85f1a41f9a" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.569474 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6mtn4" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.847600 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.908336 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kcvh\" (UniqueName: \"kubernetes.io/projected/d8f54e64-697d-41b5-83fd-295ce5143d04-kube-api-access-8kcvh\") pod \"d8f54e64-697d-41b5-83fd-295ce5143d04\" (UID: \"d8f54e64-697d-41b5-83fd-295ce5143d04\") " Oct 02 10:03:52 crc kubenswrapper[4771]: I1002 10:03:52.921051 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8f54e64-697d-41b5-83fd-295ce5143d04-kube-api-access-8kcvh" (OuterVolumeSpecName: "kube-api-access-8kcvh") pod "d8f54e64-697d-41b5-83fd-295ce5143d04" (UID: "d8f54e64-697d-41b5-83fd-295ce5143d04"). InnerVolumeSpecName "kube-api-access-8kcvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.012024 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kcvh\" (UniqueName: \"kubernetes.io/projected/d8f54e64-697d-41b5-83fd-295ce5143d04-kube-api-access-8kcvh\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.298713 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.319358 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5cfh\" (UniqueName: \"kubernetes.io/projected/e37a8032-f348-4e48-ac3a-9da319680ad2-kube-api-access-s5cfh\") pod \"e37a8032-f348-4e48-ac3a-9da319680ad2\" (UID: \"e37a8032-f348-4e48-ac3a-9da319680ad2\") " Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.319759 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e37a8032-f348-4e48-ac3a-9da319680ad2-config-data\") pod \"e37a8032-f348-4e48-ac3a-9da319680ad2\" (UID: \"e37a8032-f348-4e48-ac3a-9da319680ad2\") " Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.320239 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37a8032-f348-4e48-ac3a-9da319680ad2-combined-ca-bundle\") pod \"e37a8032-f348-4e48-ac3a-9da319680ad2\" (UID: \"e37a8032-f348-4e48-ac3a-9da319680ad2\") " Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.356099 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e37a8032-f348-4e48-ac3a-9da319680ad2-kube-api-access-s5cfh" (OuterVolumeSpecName: "kube-api-access-s5cfh") pod "e37a8032-f348-4e48-ac3a-9da319680ad2" (UID: "e37a8032-f348-4e48-ac3a-9da319680ad2"). InnerVolumeSpecName "kube-api-access-s5cfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:53 crc kubenswrapper[4771]: E1002 10:03:53.400767 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285 is running failed: container process not found" containerID="2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:03:53 crc kubenswrapper[4771]: E1002 10:03:53.413072 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285 is running failed: container process not found" containerID="2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:03:53 crc kubenswrapper[4771]: E1002 10:03:53.416896 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285 is running failed: container process not found" containerID="2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:03:53 crc kubenswrapper[4771]: E1002 10:03:53.416997 4771 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ff529351-3126-40ed-9d60-87959d520219" containerName="nova-scheduler-scheduler" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.440835 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5cfh\" (UniqueName: \"kubernetes.io/projected/e37a8032-f348-4e48-ac3a-9da319680ad2-kube-api-access-s5cfh\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.472799 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.548345 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37a8032-f348-4e48-ac3a-9da319680ad2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e37a8032-f348-4e48-ac3a-9da319680ad2" (UID: "e37a8032-f348-4e48-ac3a-9da319680ad2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.557621 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37a8032-f348-4e48-ac3a-9da319680ad2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.573390 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 02 10:03:53 crc kubenswrapper[4771]: E1002 10:03:53.574102 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f54e64-697d-41b5-83fd-295ce5143d04" containerName="kube-state-metrics" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.574233 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f54e64-697d-41b5-83fd-295ce5143d04" containerName="kube-state-metrics" Oct 02 10:03:53 crc kubenswrapper[4771]: E1002 10:03:53.574273 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d8b4155-64a1-4c45-b368-72013bcdb7aa" containerName="aodh-db-sync" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.574280 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d8b4155-64a1-4c45-b368-72013bcdb7aa" containerName="aodh-db-sync" Oct 02 10:03:53 crc kubenswrapper[4771]: E1002 10:03:53.574304 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37a8032-f348-4e48-ac3a-9da319680ad2" containerName="mysqld-exporter" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.574310 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37a8032-f348-4e48-ac3a-9da319680ad2" containerName="mysqld-exporter" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.575047 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="e37a8032-f348-4e48-ac3a-9da319680ad2" containerName="mysqld-exporter" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.575067 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8f54e64-697d-41b5-83fd-295ce5143d04" containerName="kube-state-metrics" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.575093 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d8b4155-64a1-4c45-b368-72013bcdb7aa" containerName="aodh-db-sync" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.578325 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.584947 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.586668 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.587016 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-mnlz2" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.590528 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.596746 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37a8032-f348-4e48-ac3a-9da319680ad2-config-data" (OuterVolumeSpecName: "config-data") pod "e37a8032-f348-4e48-ac3a-9da319680ad2" (UID: "e37a8032-f348-4e48-ac3a-9da319680ad2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.635685 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"e37a8032-f348-4e48-ac3a-9da319680ad2","Type":"ContainerDied","Data":"d8e0fd6c12260c19548e5f5e34613f500735092853fc6e1cf9604725264ed6fe"} Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.635747 4771 scope.go:117] "RemoveContainer" containerID="c0e9c71d39ee3db61f689d3ee4d53ba4357ca3415bb3eb88f0392121dbb5f8b6" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.635852 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.668550 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " pod="openstack/aodh-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.668742 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-scripts\") pod \"aodh-0\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " pod="openstack/aodh-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.668915 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-config-data\") pod \"aodh-0\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " pod="openstack/aodh-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.670566 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.670581 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d8f54e64-697d-41b5-83fd-295ce5143d04","Type":"ContainerDied","Data":"a23f78a1259dbb38e2e04485fa46cc26df896546b85ced9b3773c58baa2d17e3"} Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.675275 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r5xh\" (UniqueName: \"kubernetes.io/projected/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-kube-api-access-6r5xh\") pod \"aodh-0\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " pod="openstack/aodh-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.675589 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e37a8032-f348-4e48-ac3a-9da319680ad2-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.698689 4771 generic.go:334] "Generic (PLEG): container finished" podID="ff529351-3126-40ed-9d60-87959d520219" containerID="2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285" exitCode=0 Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.782885 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r5xh\" (UniqueName: \"kubernetes.io/projected/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-kube-api-access-6r5xh\") pod \"aodh-0\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " pod="openstack/aodh-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.783007 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " pod="openstack/aodh-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.783074 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-scripts\") pod \"aodh-0\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " pod="openstack/aodh-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.783157 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-config-data\") pod \"aodh-0\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " pod="openstack/aodh-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.791515 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " pod="openstack/aodh-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.800892 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-scripts\") pod \"aodh-0\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " pod="openstack/aodh-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.801452 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-config-data\") pod \"aodh-0\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " pod="openstack/aodh-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.813773 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r5xh\" (UniqueName: \"kubernetes.io/projected/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-kube-api-access-6r5xh\") pod \"aodh-0\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " pod="openstack/aodh-0" Oct 02 10:03:53 crc kubenswrapper[4771]: I1002 10:03:53.939390 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.046767 4771 scope.go:117] "RemoveContainer" containerID="d08794f587f9bdf237349764132dd957e40260917b09698e06252abe9c0e6ae9" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.107900 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.126462 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81f80e00-ac5e-4675-b349-c2aac3bd66ca" path="/var/lib/kubelet/pods/81f80e00-ac5e-4675-b349-c2aac3bd66ca/volumes" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.138364 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.138407 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78504a04-3786-414a-94f9-e7cd45d2deee","Type":"ContainerStarted","Data":"807be8f229569cbc1cd7fff6b0f5458a1633c2b343c1e1b0310d6757c31557f8"} Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.138440 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ff529351-3126-40ed-9d60-87959d520219","Type":"ContainerDied","Data":"2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285"} Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.138464 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.161822 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Oct 02 10:03:54 crc kubenswrapper[4771]: E1002 10:03:54.162673 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff529351-3126-40ed-9d60-87959d520219" containerName="nova-scheduler-scheduler" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.162719 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff529351-3126-40ed-9d60-87959d520219" containerName="nova-scheduler-scheduler" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.163147 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff529351-3126-40ed-9d60-87959d520219" containerName="nova-scheduler-scheduler" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.164266 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.168457 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.171461 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.171782 4771 scope.go:117] "RemoveContainer" containerID="2636ba554b76c426382114a1c3112e05f8b07e5168ff1acaf4c1ca45294cd285" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.214667 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.217113 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-config-data\") pod \"ff529351-3126-40ed-9d60-87959d520219\" (UID: \"ff529351-3126-40ed-9d60-87959d520219\") " Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.217422 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-combined-ca-bundle\") pod \"ff529351-3126-40ed-9d60-87959d520219\" (UID: \"ff529351-3126-40ed-9d60-87959d520219\") " Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.217476 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msmrn\" (UniqueName: \"kubernetes.io/projected/ff529351-3126-40ed-9d60-87959d520219-kube-api-access-msmrn\") pod \"ff529351-3126-40ed-9d60-87959d520219\" (UID: \"ff529351-3126-40ed-9d60-87959d520219\") " Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.219446 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpqqd\" (UniqueName: \"kubernetes.io/projected/f57f79e3-d1a8-4830-a50b-6d414f041ec4-kube-api-access-kpqqd\") pod \"mysqld-exporter-0\" (UID: \"f57f79e3-d1a8-4830-a50b-6d414f041ec4\") " pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.220714 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/f57f79e3-d1a8-4830-a50b-6d414f041ec4-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"f57f79e3-d1a8-4830-a50b-6d414f041ec4\") " pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.220838 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f57f79e3-d1a8-4830-a50b-6d414f041ec4-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"f57f79e3-d1a8-4830-a50b-6d414f041ec4\") " pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.220918 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f57f79e3-d1a8-4830-a50b-6d414f041ec4-config-data\") pod \"mysqld-exporter-0\" (UID: \"f57f79e3-d1a8-4830-a50b-6d414f041ec4\") " pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.249153 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff529351-3126-40ed-9d60-87959d520219-kube-api-access-msmrn" (OuterVolumeSpecName: "kube-api-access-msmrn") pod "ff529351-3126-40ed-9d60-87959d520219" (UID: "ff529351-3126-40ed-9d60-87959d520219"). InnerVolumeSpecName "kube-api-access-msmrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.325707 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpqqd\" (UniqueName: \"kubernetes.io/projected/f57f79e3-d1a8-4830-a50b-6d414f041ec4-kube-api-access-kpqqd\") pod \"mysqld-exporter-0\" (UID: \"f57f79e3-d1a8-4830-a50b-6d414f041ec4\") " pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.333271 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/f57f79e3-d1a8-4830-a50b-6d414f041ec4-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"f57f79e3-d1a8-4830-a50b-6d414f041ec4\") " pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.333453 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f57f79e3-d1a8-4830-a50b-6d414f041ec4-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"f57f79e3-d1a8-4830-a50b-6d414f041ec4\") " pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.333583 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f57f79e3-d1a8-4830-a50b-6d414f041ec4-config-data\") pod \"mysqld-exporter-0\" (UID: \"f57f79e3-d1a8-4830-a50b-6d414f041ec4\") " pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.334043 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msmrn\" (UniqueName: \"kubernetes.io/projected/ff529351-3126-40ed-9d60-87959d520219-kube-api-access-msmrn\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.349786 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f57f79e3-d1a8-4830-a50b-6d414f041ec4-config-data\") pod \"mysqld-exporter-0\" (UID: \"f57f79e3-d1a8-4830-a50b-6d414f041ec4\") " pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.351659 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f57f79e3-d1a8-4830-a50b-6d414f041ec4-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"f57f79e3-d1a8-4830-a50b-6d414f041ec4\") " pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.361685 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpqqd\" (UniqueName: \"kubernetes.io/projected/f57f79e3-d1a8-4830-a50b-6d414f041ec4-kube-api-access-kpqqd\") pod \"mysqld-exporter-0\" (UID: \"f57f79e3-d1a8-4830-a50b-6d414f041ec4\") " pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.370872 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.378426 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/f57f79e3-d1a8-4830-a50b-6d414f041ec4-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"f57f79e3-d1a8-4830-a50b-6d414f041ec4\") " pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.396055 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.408305 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.418858 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.421876 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.423493 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.438831 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:03:54 crc kubenswrapper[4771]: E1002 10:03:54.439929 4771 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-combined-ca-bundle podName:ff529351-3126-40ed-9d60-87959d520219 nodeName:}" failed. No retries permitted until 2025-10-02 10:03:54.93989918 +0000 UTC m=+1622.587584247 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-combined-ca-bundle") pod "ff529351-3126-40ed-9d60-87959d520219" (UID: "ff529351-3126-40ed-9d60-87959d520219") : error deleting /var/lib/kubelet/pods/ff529351-3126-40ed-9d60-87959d520219/volume-subpaths: remove /var/lib/kubelet/pods/ff529351-3126-40ed-9d60-87959d520219/volume-subpaths: no such file or directory Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.450740 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-config-data" (OuterVolumeSpecName: "config-data") pod "ff529351-3126-40ed-9d60-87959d520219" (UID: "ff529351-3126-40ed-9d60-87959d520219"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.497468 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.541176 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9srdt\" (UniqueName: \"kubernetes.io/projected/b20929e5-d840-461a-be05-c9f2783540f5-kube-api-access-9srdt\") pod \"kube-state-metrics-0\" (UID: \"b20929e5-d840-461a-be05-c9f2783540f5\") " pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.548550 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b20929e5-d840-461a-be05-c9f2783540f5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b20929e5-d840-461a-be05-c9f2783540f5\") " pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.548660 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b20929e5-d840-461a-be05-c9f2783540f5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b20929e5-d840-461a-be05-c9f2783540f5\") " pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.548863 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b20929e5-d840-461a-be05-c9f2783540f5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b20929e5-d840-461a-be05-c9f2783540f5\") " pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.549053 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.651199 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9srdt\" (UniqueName: \"kubernetes.io/projected/b20929e5-d840-461a-be05-c9f2783540f5-kube-api-access-9srdt\") pod \"kube-state-metrics-0\" (UID: \"b20929e5-d840-461a-be05-c9f2783540f5\") " pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.651763 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b20929e5-d840-461a-be05-c9f2783540f5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b20929e5-d840-461a-be05-c9f2783540f5\") " pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.651820 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b20929e5-d840-461a-be05-c9f2783540f5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b20929e5-d840-461a-be05-c9f2783540f5\") " pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.651940 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b20929e5-d840-461a-be05-c9f2783540f5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b20929e5-d840-461a-be05-c9f2783540f5\") " pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.656883 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b20929e5-d840-461a-be05-c9f2783540f5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b20929e5-d840-461a-be05-c9f2783540f5\") " pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.657619 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b20929e5-d840-461a-be05-c9f2783540f5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b20929e5-d840-461a-be05-c9f2783540f5\") " pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.659303 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b20929e5-d840-461a-be05-c9f2783540f5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b20929e5-d840-461a-be05-c9f2783540f5\") " pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.682741 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9srdt\" (UniqueName: \"kubernetes.io/projected/b20929e5-d840-461a-be05-c9f2783540f5-kube-api-access-9srdt\") pod \"kube-state-metrics-0\" (UID: \"b20929e5-d840-461a-be05-c9f2783540f5\") " pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.754843 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ff529351-3126-40ed-9d60-87959d520219","Type":"ContainerDied","Data":"abcc81ff8aac70a133b6b5adb36daf0b68556f0e59a78f41ba913285f95ceee0"} Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.755050 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.786661 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78504a04-3786-414a-94f9-e7cd45d2deee","Type":"ContainerStarted","Data":"91f1ccf5a30da789ef16da87b41d107f570e92baafa4e58ff49db615a66e6e1d"} Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.908119 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.960243 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-combined-ca-bundle\") pod \"ff529351-3126-40ed-9d60-87959d520219\" (UID: \"ff529351-3126-40ed-9d60-87959d520219\") " Oct 02 10:03:54 crc kubenswrapper[4771]: I1002 10:03:54.971333 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff529351-3126-40ed-9d60-87959d520219" (UID: "ff529351-3126-40ed-9d60-87959d520219"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.009458 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.039727 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.068428 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff529351-3126-40ed-9d60-87959d520219-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.157457 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.192886 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.219370 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.250067 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.250556 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.257620 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.375826 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-config-data\") pod \"nova-scheduler-0\" (UID: \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.375963 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.376048 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nwx9\" (UniqueName: \"kubernetes.io/projected/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-kube-api-access-8nwx9\") pod \"nova-scheduler-0\" (UID: \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.404592 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.479082 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nwx9\" (UniqueName: \"kubernetes.io/projected/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-kube-api-access-8nwx9\") pod \"nova-scheduler-0\" (UID: \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.479303 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-config-data\") pod \"nova-scheduler-0\" (UID: \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.479444 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.489987 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-config-data\") pod \"nova-scheduler-0\" (UID: \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.490918 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.530025 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nwx9\" (UniqueName: \"kubernetes.io/projected/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-kube-api-access-8nwx9\") pod \"nova-scheduler-0\" (UID: \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\") " pod="openstack/nova-scheduler-0" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.897072 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8f54e64-697d-41b5-83fd-295ce5143d04" path="/var/lib/kubelet/pods/d8f54e64-697d-41b5-83fd-295ce5143d04/volumes" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.899519 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e37a8032-f348-4e48-ac3a-9da319680ad2" path="/var/lib/kubelet/pods/e37a8032-f348-4e48-ac3a-9da319680ad2/volumes" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.902756 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff529351-3126-40ed-9d60-87959d520219" path="/var/lib/kubelet/pods/ff529351-3126-40ed-9d60-87959d520219/volumes" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.910389 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.924586 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78504a04-3786-414a-94f9-e7cd45d2deee","Type":"ContainerStarted","Data":"ad62c8eb290c27c2af0e52f0b674e9aee22b87f31fbab7a32dfdf97e467f834e"} Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.945121 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1","Type":"ContainerStarted","Data":"11bfaed686721cc97fe9e21fe0d67e4d51e2a58f8c2758858cce3c71b1df0022"} Oct 02 10:03:55 crc kubenswrapper[4771]: I1002 10:03:55.984888 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.984861138 podStartE2EDuration="4.984861138s" podCreationTimestamp="2025-10-02 10:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:03:55.976614624 +0000 UTC m=+1623.624299691" watchObservedRunningTime="2025-10-02 10:03:55.984861138 +0000 UTC m=+1623.632546205" Oct 02 10:03:56 crc kubenswrapper[4771]: I1002 10:03:56.152002 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:03:56 crc kubenswrapper[4771]: I1002 10:03:56.812827 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:03:56 crc kubenswrapper[4771]: I1002 10:03:56.987273 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1","Type":"ContainerStarted","Data":"78e04f9546fe73ae805651aa40ff364d3723d2decf17de1b2cba0a143d57cd81"} Oct 02 10:03:56 crc kubenswrapper[4771]: I1002 10:03:56.987455 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 02 10:03:56 crc kubenswrapper[4771]: I1002 10:03:56.996476 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"f57f79e3-d1a8-4830-a50b-6d414f041ec4","Type":"ContainerStarted","Data":"c4fb3c6f718944ea873de06b7f6510d91c99b075a354e299d51e3ad3d5a42587"} Oct 02 10:03:57 crc kubenswrapper[4771]: I1002 10:03:57.023284 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d343eb2f-faff-47b8-aad6-9d24a3fc39c7","Type":"ContainerStarted","Data":"b5e9ee1f4991d03e79d5ff8918074a64c44b2359ea7671b5765eb36f46c043bd"} Oct 02 10:03:57 crc kubenswrapper[4771]: I1002 10:03:57.031016 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b20929e5-d840-461a-be05-c9f2783540f5","Type":"ContainerStarted","Data":"14e82e90573a8c80f32e20f4cf40a8fbf0c0593f98e5a7547dc32c02e7741271"} Oct 02 10:03:57 crc kubenswrapper[4771]: I1002 10:03:57.842346 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:03:57 crc kubenswrapper[4771]: I1002 10:03:57.843947 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="ceilometer-central-agent" containerID="cri-o://97fdbd831c00862cda41bbf77db00d97cea8c8d22204850b73ed21a0c5975289" gracePeriod=30 Oct 02 10:03:57 crc kubenswrapper[4771]: I1002 10:03:57.844118 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="sg-core" containerID="cri-o://da6a2166316f9f1ead80c45c13320c5f8cdf387ddf0ee787ebba09e521d4cf4e" gracePeriod=30 Oct 02 10:03:57 crc kubenswrapper[4771]: I1002 10:03:57.844196 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="ceilometer-notification-agent" containerID="cri-o://9467df9eaffc2ba3c84b6bce0d755e29d56809a3fee62731a1e3e535a9af9035" gracePeriod=30 Oct 02 10:03:57 crc kubenswrapper[4771]: I1002 10:03:57.844171 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="proxy-httpd" containerID="cri-o://d708a48848c28ecb797ca149a414b0ed2e22f15e3d4c604bdf2da2ca2b780207" gracePeriod=30 Oct 02 10:03:58 crc kubenswrapper[4771]: I1002 10:03:58.064435 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"f57f79e3-d1a8-4830-a50b-6d414f041ec4","Type":"ContainerStarted","Data":"9efeb8ef8ff2bdfa47b81afa4d3ab55603a324cb055a36e1bc8ede64f085f215"} Oct 02 10:03:58 crc kubenswrapper[4771]: I1002 10:03:58.092446 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.968081222 podStartE2EDuration="4.09242046s" podCreationTimestamp="2025-10-02 10:03:54 +0000 UTC" firstStartedPulling="2025-10-02 10:03:56.014175182 +0000 UTC m=+1623.661860249" lastFinishedPulling="2025-10-02 10:03:57.13851442 +0000 UTC m=+1624.786199487" observedRunningTime="2025-10-02 10:03:58.086388138 +0000 UTC m=+1625.734073205" watchObservedRunningTime="2025-10-02 10:03:58.09242046 +0000 UTC m=+1625.740105537" Oct 02 10:03:58 crc kubenswrapper[4771]: I1002 10:03:58.094814 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d343eb2f-faff-47b8-aad6-9d24a3fc39c7","Type":"ContainerStarted","Data":"eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911"} Oct 02 10:03:58 crc kubenswrapper[4771]: I1002 10:03:58.109402 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b20929e5-d840-461a-be05-c9f2783540f5","Type":"ContainerStarted","Data":"c7d196a8d35e2ebfbe926181b860bd4d3fd0698570535e625d7844deff2b59cd"} Oct 02 10:03:58 crc kubenswrapper[4771]: I1002 10:03:58.110723 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 10:03:58 crc kubenswrapper[4771]: I1002 10:03:58.127894 4771 generic.go:334] "Generic (PLEG): container finished" podID="9158fae1-e129-47d9-af54-2105c7d79e41" containerID="da6a2166316f9f1ead80c45c13320c5f8cdf387ddf0ee787ebba09e521d4cf4e" exitCode=2 Oct 02 10:03:58 crc kubenswrapper[4771]: I1002 10:03:58.128331 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9158fae1-e129-47d9-af54-2105c7d79e41","Type":"ContainerDied","Data":"da6a2166316f9f1ead80c45c13320c5f8cdf387ddf0ee787ebba09e521d4cf4e"} Oct 02 10:03:58 crc kubenswrapper[4771]: I1002 10:03:58.185835 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.185805996 podStartE2EDuration="3.185805996s" podCreationTimestamp="2025-10-02 10:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:03:58.131518592 +0000 UTC m=+1625.779203659" watchObservedRunningTime="2025-10-02 10:03:58.185805996 +0000 UTC m=+1625.833491063" Oct 02 10:03:58 crc kubenswrapper[4771]: I1002 10:03:58.250486 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.817968395 podStartE2EDuration="4.250459305s" podCreationTimestamp="2025-10-02 10:03:54 +0000 UTC" firstStartedPulling="2025-10-02 10:03:56.144467868 +0000 UTC m=+1623.792152935" lastFinishedPulling="2025-10-02 10:03:56.576958788 +0000 UTC m=+1624.224643845" observedRunningTime="2025-10-02 10:03:58.169949895 +0000 UTC m=+1625.817634972" watchObservedRunningTime="2025-10-02 10:03:58.250459305 +0000 UTC m=+1625.898144372" Oct 02 10:03:59 crc kubenswrapper[4771]: I1002 10:03:59.168189 4771 generic.go:334] "Generic (PLEG): container finished" podID="9158fae1-e129-47d9-af54-2105c7d79e41" containerID="d708a48848c28ecb797ca149a414b0ed2e22f15e3d4c604bdf2da2ca2b780207" exitCode=0 Oct 02 10:03:59 crc kubenswrapper[4771]: I1002 10:03:59.168672 4771 generic.go:334] "Generic (PLEG): container finished" podID="9158fae1-e129-47d9-af54-2105c7d79e41" containerID="97fdbd831c00862cda41bbf77db00d97cea8c8d22204850b73ed21a0c5975289" exitCode=0 Oct 02 10:03:59 crc kubenswrapper[4771]: I1002 10:03:59.168270 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9158fae1-e129-47d9-af54-2105c7d79e41","Type":"ContainerDied","Data":"d708a48848c28ecb797ca149a414b0ed2e22f15e3d4c604bdf2da2ca2b780207"} Oct 02 10:03:59 crc kubenswrapper[4771]: I1002 10:03:59.170300 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9158fae1-e129-47d9-af54-2105c7d79e41","Type":"ContainerDied","Data":"97fdbd831c00862cda41bbf77db00d97cea8c8d22204850b73ed21a0c5975289"} Oct 02 10:03:59 crc kubenswrapper[4771]: I1002 10:03:59.719847 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.201307 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1","Type":"ContainerStarted","Data":"01aeca9e1ffa450c46ed0ded04944aea8c92057279c8ac0632ac8e8fdcd1f883"} Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.552538 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q8kv5"] Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.557505 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.588891 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lt6c\" (UniqueName: \"kubernetes.io/projected/8072221f-4513-4f85-b4ea-f06d0b81214c-kube-api-access-4lt6c\") pod \"redhat-marketplace-q8kv5\" (UID: \"8072221f-4513-4f85-b4ea-f06d0b81214c\") " pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.589274 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8072221f-4513-4f85-b4ea-f06d0b81214c-utilities\") pod \"redhat-marketplace-q8kv5\" (UID: \"8072221f-4513-4f85-b4ea-f06d0b81214c\") " pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.589350 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8072221f-4513-4f85-b4ea-f06d0b81214c-catalog-content\") pod \"redhat-marketplace-q8kv5\" (UID: \"8072221f-4513-4f85-b4ea-f06d0b81214c\") " pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.614300 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8kv5"] Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.682469 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:04:00 crc kubenswrapper[4771]: E1002 10:04:00.684885 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.693376 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lt6c\" (UniqueName: \"kubernetes.io/projected/8072221f-4513-4f85-b4ea-f06d0b81214c-kube-api-access-4lt6c\") pod \"redhat-marketplace-q8kv5\" (UID: \"8072221f-4513-4f85-b4ea-f06d0b81214c\") " pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.693672 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8072221f-4513-4f85-b4ea-f06d0b81214c-utilities\") pod \"redhat-marketplace-q8kv5\" (UID: \"8072221f-4513-4f85-b4ea-f06d0b81214c\") " pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.693729 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8072221f-4513-4f85-b4ea-f06d0b81214c-catalog-content\") pod \"redhat-marketplace-q8kv5\" (UID: \"8072221f-4513-4f85-b4ea-f06d0b81214c\") " pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.694643 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8072221f-4513-4f85-b4ea-f06d0b81214c-catalog-content\") pod \"redhat-marketplace-q8kv5\" (UID: \"8072221f-4513-4f85-b4ea-f06d0b81214c\") " pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.697518 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8072221f-4513-4f85-b4ea-f06d0b81214c-utilities\") pod \"redhat-marketplace-q8kv5\" (UID: \"8072221f-4513-4f85-b4ea-f06d0b81214c\") " pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.743098 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lt6c\" (UniqueName: \"kubernetes.io/projected/8072221f-4513-4f85-b4ea-f06d0b81214c-kube-api-access-4lt6c\") pod \"redhat-marketplace-q8kv5\" (UID: \"8072221f-4513-4f85-b4ea-f06d0b81214c\") " pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.909117 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:00 crc kubenswrapper[4771]: I1002 10:04:00.911003 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.172760 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8kv5"] Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.252819 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.252856 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.260816 4771 generic.go:334] "Generic (PLEG): container finished" podID="9158fae1-e129-47d9-af54-2105c7d79e41" containerID="9467df9eaffc2ba3c84b6bce0d755e29d56809a3fee62731a1e3e535a9af9035" exitCode=0 Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.260858 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9158fae1-e129-47d9-af54-2105c7d79e41","Type":"ContainerDied","Data":"9467df9eaffc2ba3c84b6bce0d755e29d56809a3fee62731a1e3e535a9af9035"} Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.264292 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1","Type":"ContainerStarted","Data":"7a4a5dcedcdd2808566d1dece6c391df5b560ba0717684ab2acce35ea635a5d1"} Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.269295 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8kv5" event={"ID":"8072221f-4513-4f85-b4ea-f06d0b81214c","Type":"ContainerStarted","Data":"d5af6363beb94a27d4291c0997172881647c08e567470c80e59369ce5b714b99"} Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.870694 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.963074 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmlpr\" (UniqueName: \"kubernetes.io/projected/9158fae1-e129-47d9-af54-2105c7d79e41-kube-api-access-cmlpr\") pod \"9158fae1-e129-47d9-af54-2105c7d79e41\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.963135 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9158fae1-e129-47d9-af54-2105c7d79e41-log-httpd\") pod \"9158fae1-e129-47d9-af54-2105c7d79e41\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.963244 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9158fae1-e129-47d9-af54-2105c7d79e41-run-httpd\") pod \"9158fae1-e129-47d9-af54-2105c7d79e41\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.963314 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-config-data\") pod \"9158fae1-e129-47d9-af54-2105c7d79e41\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.963357 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-combined-ca-bundle\") pod \"9158fae1-e129-47d9-af54-2105c7d79e41\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.963506 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-sg-core-conf-yaml\") pod \"9158fae1-e129-47d9-af54-2105c7d79e41\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.964418 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-scripts\") pod \"9158fae1-e129-47d9-af54-2105c7d79e41\" (UID: \"9158fae1-e129-47d9-af54-2105c7d79e41\") " Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.965791 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9158fae1-e129-47d9-af54-2105c7d79e41-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9158fae1-e129-47d9-af54-2105c7d79e41" (UID: "9158fae1-e129-47d9-af54-2105c7d79e41"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.966028 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9158fae1-e129-47d9-af54-2105c7d79e41-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9158fae1-e129-47d9-af54-2105c7d79e41" (UID: "9158fae1-e129-47d9-af54-2105c7d79e41"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.979347 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-scripts" (OuterVolumeSpecName: "scripts") pod "9158fae1-e129-47d9-af54-2105c7d79e41" (UID: "9158fae1-e129-47d9-af54-2105c7d79e41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:02 crc kubenswrapper[4771]: I1002 10:04:02.979445 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9158fae1-e129-47d9-af54-2105c7d79e41-kube-api-access-cmlpr" (OuterVolumeSpecName: "kube-api-access-cmlpr") pod "9158fae1-e129-47d9-af54-2105c7d79e41" (UID: "9158fae1-e129-47d9-af54-2105c7d79e41"). InnerVolumeSpecName "kube-api-access-cmlpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.058830 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9158fae1-e129-47d9-af54-2105c7d79e41" (UID: "9158fae1-e129-47d9-af54-2105c7d79e41"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.067979 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.068019 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmlpr\" (UniqueName: \"kubernetes.io/projected/9158fae1-e129-47d9-af54-2105c7d79e41-kube-api-access-cmlpr\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.068030 4771 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9158fae1-e129-47d9-af54-2105c7d79e41-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.068041 4771 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9158fae1-e129-47d9-af54-2105c7d79e41-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.068049 4771 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.179401 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9158fae1-e129-47d9-af54-2105c7d79e41" (UID: "9158fae1-e129-47d9-af54-2105c7d79e41"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.266399 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-config-data" (OuterVolumeSpecName: "config-data") pod "9158fae1-e129-47d9-af54-2105c7d79e41" (UID: "9158fae1-e129-47d9-af54-2105c7d79e41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.274457 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.274498 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9158fae1-e129-47d9-af54-2105c7d79e41-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.313706 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9158fae1-e129-47d9-af54-2105c7d79e41","Type":"ContainerDied","Data":"7f5f125117d1b1388a8503a71ec4c602610220b192068f6624d985dd54031912"} Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.313776 4771 scope.go:117] "RemoveContainer" containerID="d708a48848c28ecb797ca149a414b0ed2e22f15e3d4c604bdf2da2ca2b780207" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.314039 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.323726 4771 generic.go:334] "Generic (PLEG): container finished" podID="8072221f-4513-4f85-b4ea-f06d0b81214c" containerID="b643df3d0efcea91de0fc7906e4b194ded0b5e1777bf0eaf10e4026496870509" exitCode=0 Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.323773 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8kv5" event={"ID":"8072221f-4513-4f85-b4ea-f06d0b81214c","Type":"ContainerDied","Data":"b643df3d0efcea91de0fc7906e4b194ded0b5e1777bf0eaf10e4026496870509"} Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.335782 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="78504a04-3786-414a-94f9-e7cd45d2deee" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.247:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.335978 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="78504a04-3786-414a-94f9-e7cd45d2deee" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.247:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.440757 4771 scope.go:117] "RemoveContainer" containerID="da6a2166316f9f1ead80c45c13320c5f8cdf387ddf0ee787ebba09e521d4cf4e" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.507335 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.517484 4771 scope.go:117] "RemoveContainer" containerID="9467df9eaffc2ba3c84b6bce0d755e29d56809a3fee62731a1e3e535a9af9035" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.527544 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.538904 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:04:03 crc kubenswrapper[4771]: E1002 10:04:03.540613 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="ceilometer-notification-agent" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.540646 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="ceilometer-notification-agent" Oct 02 10:04:03 crc kubenswrapper[4771]: E1002 10:04:03.540687 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="proxy-httpd" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.540698 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="proxy-httpd" Oct 02 10:04:03 crc kubenswrapper[4771]: E1002 10:04:03.540723 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="sg-core" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.540732 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="sg-core" Oct 02 10:04:03 crc kubenswrapper[4771]: E1002 10:04:03.540778 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="ceilometer-central-agent" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.540789 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="ceilometer-central-agent" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.541302 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="ceilometer-notification-agent" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.541337 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="sg-core" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.541487 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="proxy-httpd" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.541508 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" containerName="ceilometer-central-agent" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.549915 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.553005 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.554136 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.554630 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.554912 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.633887 4771 scope.go:117] "RemoveContainer" containerID="97fdbd831c00862cda41bbf77db00d97cea8c8d22204850b73ed21a0c5975289" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.711721 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9158fae1-e129-47d9-af54-2105c7d79e41" path="/var/lib/kubelet/pods/9158fae1-e129-47d9-af54-2105c7d79e41/volumes" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.720263 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9k95\" (UniqueName: \"kubernetes.io/projected/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-kube-api-access-n9k95\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.720361 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-log-httpd\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.720391 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.720544 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-config-data\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.720574 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-run-httpd\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.720594 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.720617 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-scripts\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.720843 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.823339 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-log-httpd\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.823954 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.824254 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-config-data\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.824298 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-run-httpd\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.824334 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.824373 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-scripts\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.824444 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.824539 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9k95\" (UniqueName: \"kubernetes.io/projected/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-kube-api-access-n9k95\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.833001 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-log-httpd\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.833378 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-run-httpd\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.841702 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.841713 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-scripts\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.842902 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.844643 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.845268 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9k95\" (UniqueName: \"kubernetes.io/projected/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-kube-api-access-n9k95\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.851551 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-config-data\") pod \"ceilometer-0\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " pod="openstack/ceilometer-0" Oct 02 10:04:03 crc kubenswrapper[4771]: I1002 10:04:03.926757 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:04:04 crc kubenswrapper[4771]: I1002 10:04:04.936597 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 10:04:05 crc kubenswrapper[4771]: I1002 10:04:05.109103 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:04:05 crc kubenswrapper[4771]: I1002 10:04:05.358290 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1","Type":"ContainerStarted","Data":"7f020577297a4d53ef7fdb1e0a360574e80bc0959ace2d891f409dd0403c7541"} Oct 02 10:04:05 crc kubenswrapper[4771]: I1002 10:04:05.358515 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-listener" containerID="cri-o://7f020577297a4d53ef7fdb1e0a360574e80bc0959ace2d891f409dd0403c7541" gracePeriod=30 Oct 02 10:04:05 crc kubenswrapper[4771]: I1002 10:04:05.359049 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-notifier" containerID="cri-o://7a4a5dcedcdd2808566d1dece6c391df5b560ba0717684ab2acce35ea635a5d1" gracePeriod=30 Oct 02 10:04:05 crc kubenswrapper[4771]: I1002 10:04:05.359228 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-evaluator" containerID="cri-o://01aeca9e1ffa450c46ed0ded04944aea8c92057279c8ac0632ac8e8fdcd1f883" gracePeriod=30 Oct 02 10:04:05 crc kubenswrapper[4771]: I1002 10:04:05.359346 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-api" containerID="cri-o://78e04f9546fe73ae805651aa40ff364d3723d2decf17de1b2cba0a143d57cd81" gracePeriod=30 Oct 02 10:04:05 crc kubenswrapper[4771]: I1002 10:04:05.365277 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8kv5" event={"ID":"8072221f-4513-4f85-b4ea-f06d0b81214c","Type":"ContainerStarted","Data":"e08dea8c94e8f50f4c57abd2be2bb92804b75ef06bf9fcff15d36c817675f59b"} Oct 02 10:04:05 crc kubenswrapper[4771]: I1002 10:04:05.367401 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc","Type":"ContainerStarted","Data":"03dbf26878bc3888bff3165d8550636c52c701528edddf8412ef9cd9662d4c65"} Oct 02 10:04:05 crc kubenswrapper[4771]: I1002 10:04:05.381873 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.850565651 podStartE2EDuration="12.381851658s" podCreationTimestamp="2025-10-02 10:03:53 +0000 UTC" firstStartedPulling="2025-10-02 10:03:55.039378629 +0000 UTC m=+1622.687063696" lastFinishedPulling="2025-10-02 10:04:04.570664636 +0000 UTC m=+1632.218349703" observedRunningTime="2025-10-02 10:04:05.379624544 +0000 UTC m=+1633.027309621" watchObservedRunningTime="2025-10-02 10:04:05.381851658 +0000 UTC m=+1633.029536725" Oct 02 10:04:05 crc kubenswrapper[4771]: I1002 10:04:05.911493 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 10:04:05 crc kubenswrapper[4771]: I1002 10:04:05.949463 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 10:04:06 crc kubenswrapper[4771]: I1002 10:04:06.319419 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:04:06 crc kubenswrapper[4771]: I1002 10:04:06.383573 4771 generic.go:334] "Generic (PLEG): container finished" podID="ab155770-4f5d-43cf-a234-7c9394177ef6" containerID="28d94e68497f8ab6a78649fc80178b2cdf9e7795198e38f81dfff5002f70f06a" exitCode=137 Oct 02 10:04:06 crc kubenswrapper[4771]: I1002 10:04:06.383663 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ab155770-4f5d-43cf-a234-7c9394177ef6","Type":"ContainerDied","Data":"28d94e68497f8ab6a78649fc80178b2cdf9e7795198e38f81dfff5002f70f06a"} Oct 02 10:04:06 crc kubenswrapper[4771]: I1002 10:04:06.392629 4771 generic.go:334] "Generic (PLEG): container finished" podID="8660c7e8-08fa-49fc-ab03-f0346caedce4" containerID="e96030dfcba88964dbcab13df91d670250e729241dd3b3f7f1f3596f3d1c2b8d" exitCode=137 Oct 02 10:04:06 crc kubenswrapper[4771]: I1002 10:04:06.392717 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8660c7e8-08fa-49fc-ab03-f0346caedce4","Type":"ContainerDied","Data":"e96030dfcba88964dbcab13df91d670250e729241dd3b3f7f1f3596f3d1c2b8d"} Oct 02 10:04:06 crc kubenswrapper[4771]: E1002 10:04:06.403536 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8f54e64_697d_41b5_83fd_295ce5143d04.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8f54e64_697d_41b5_83fd_295ce5143d04.slice/crio-a23f78a1259dbb38e2e04485fa46cc26df896546b85ced9b3773c58baa2d17e3\": RecentStats: unable to find data in memory cache]" Oct 02 10:04:06 crc kubenswrapper[4771]: I1002 10:04:06.408083 4771 generic.go:334] "Generic (PLEG): container finished" podID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerID="01aeca9e1ffa450c46ed0ded04944aea8c92057279c8ac0632ac8e8fdcd1f883" exitCode=0 Oct 02 10:04:06 crc kubenswrapper[4771]: I1002 10:04:06.408114 4771 generic.go:334] "Generic (PLEG): container finished" podID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerID="78e04f9546fe73ae805651aa40ff364d3723d2decf17de1b2cba0a143d57cd81" exitCode=0 Oct 02 10:04:06 crc kubenswrapper[4771]: I1002 10:04:06.408175 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1","Type":"ContainerDied","Data":"01aeca9e1ffa450c46ed0ded04944aea8c92057279c8ac0632ac8e8fdcd1f883"} Oct 02 10:04:06 crc kubenswrapper[4771]: I1002 10:04:06.408242 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1","Type":"ContainerDied","Data":"78e04f9546fe73ae805651aa40ff364d3723d2decf17de1b2cba0a143d57cd81"} Oct 02 10:04:06 crc kubenswrapper[4771]: I1002 10:04:06.420950 4771 generic.go:334] "Generic (PLEG): container finished" podID="8072221f-4513-4f85-b4ea-f06d0b81214c" containerID="e08dea8c94e8f50f4c57abd2be2bb92804b75ef06bf9fcff15d36c817675f59b" exitCode=0 Oct 02 10:04:06 crc kubenswrapper[4771]: I1002 10:04:06.421222 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8kv5" event={"ID":"8072221f-4513-4f85-b4ea-f06d0b81214c","Type":"ContainerDied","Data":"e08dea8c94e8f50f4c57abd2be2bb92804b75ef06bf9fcff15d36c817675f59b"} Oct 02 10:04:06 crc kubenswrapper[4771]: I1002 10:04:06.485833 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.165418 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.320410 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8660c7e8-08fa-49fc-ab03-f0346caedce4-logs\") pod \"8660c7e8-08fa-49fc-ab03-f0346caedce4\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.320572 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v76hk\" (UniqueName: \"kubernetes.io/projected/8660c7e8-08fa-49fc-ab03-f0346caedce4-kube-api-access-v76hk\") pod \"8660c7e8-08fa-49fc-ab03-f0346caedce4\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.320637 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8660c7e8-08fa-49fc-ab03-f0346caedce4-combined-ca-bundle\") pod \"8660c7e8-08fa-49fc-ab03-f0346caedce4\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.321155 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8660c7e8-08fa-49fc-ab03-f0346caedce4-config-data\") pod \"8660c7e8-08fa-49fc-ab03-f0346caedce4\" (UID: \"8660c7e8-08fa-49fc-ab03-f0346caedce4\") " Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.325291 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8660c7e8-08fa-49fc-ab03-f0346caedce4-logs" (OuterVolumeSpecName: "logs") pod "8660c7e8-08fa-49fc-ab03-f0346caedce4" (UID: "8660c7e8-08fa-49fc-ab03-f0346caedce4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.339871 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8660c7e8-08fa-49fc-ab03-f0346caedce4-kube-api-access-v76hk" (OuterVolumeSpecName: "kube-api-access-v76hk") pod "8660c7e8-08fa-49fc-ab03-f0346caedce4" (UID: "8660c7e8-08fa-49fc-ab03-f0346caedce4"). InnerVolumeSpecName "kube-api-access-v76hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.372063 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8660c7e8-08fa-49fc-ab03-f0346caedce4-config-data" (OuterVolumeSpecName: "config-data") pod "8660c7e8-08fa-49fc-ab03-f0346caedce4" (UID: "8660c7e8-08fa-49fc-ab03-f0346caedce4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.413852 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8660c7e8-08fa-49fc-ab03-f0346caedce4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8660c7e8-08fa-49fc-ab03-f0346caedce4" (UID: "8660c7e8-08fa-49fc-ab03-f0346caedce4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.430572 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8660c7e8-08fa-49fc-ab03-f0346caedce4-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.430615 4771 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8660c7e8-08fa-49fc-ab03-f0346caedce4-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.430631 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v76hk\" (UniqueName: \"kubernetes.io/projected/8660c7e8-08fa-49fc-ab03-f0346caedce4-kube-api-access-v76hk\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.430644 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8660c7e8-08fa-49fc-ab03-f0346caedce4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.431730 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.443172 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8660c7e8-08fa-49fc-ab03-f0346caedce4","Type":"ContainerDied","Data":"c420ac02ef9d8cc7bda54008ec6740a952e4643409b0152bcfd2db0d296ab618"} Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.443235 4771 scope.go:117] "RemoveContainer" containerID="e96030dfcba88964dbcab13df91d670250e729241dd3b3f7f1f3596f3d1c2b8d" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.455393 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.459923 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc","Type":"ContainerStarted","Data":"3c1f6add4476e5f41d38ed17aa09b9cff2d9afe4a31f03a8086de9fdbf88837f"} Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.480842 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.481761 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ab155770-4f5d-43cf-a234-7c9394177ef6","Type":"ContainerDied","Data":"b199e517f6494b7058f19acf4963902dc80ffe473451cfe0c5b704e07c42c3bf"} Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.531862 4771 scope.go:117] "RemoveContainer" containerID="ff7920741eb56baa8292368ff934ac028cb877c2d4fc8aa93fcd48c6a1af15fb" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.533879 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrg9h\" (UniqueName: \"kubernetes.io/projected/ab155770-4f5d-43cf-a234-7c9394177ef6-kube-api-access-vrg9h\") pod \"ab155770-4f5d-43cf-a234-7c9394177ef6\" (UID: \"ab155770-4f5d-43cf-a234-7c9394177ef6\") " Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.533974 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab155770-4f5d-43cf-a234-7c9394177ef6-config-data\") pod \"ab155770-4f5d-43cf-a234-7c9394177ef6\" (UID: \"ab155770-4f5d-43cf-a234-7c9394177ef6\") " Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.534108 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab155770-4f5d-43cf-a234-7c9394177ef6-combined-ca-bundle\") pod \"ab155770-4f5d-43cf-a234-7c9394177ef6\" (UID: \"ab155770-4f5d-43cf-a234-7c9394177ef6\") " Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.540835 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab155770-4f5d-43cf-a234-7c9394177ef6-kube-api-access-vrg9h" (OuterVolumeSpecName: "kube-api-access-vrg9h") pod "ab155770-4f5d-43cf-a234-7c9394177ef6" (UID: "ab155770-4f5d-43cf-a234-7c9394177ef6"). InnerVolumeSpecName "kube-api-access-vrg9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.572143 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.603506 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab155770-4f5d-43cf-a234-7c9394177ef6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab155770-4f5d-43cf-a234-7c9394177ef6" (UID: "ab155770-4f5d-43cf-a234-7c9394177ef6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.652813 4771 scope.go:117] "RemoveContainer" containerID="28d94e68497f8ab6a78649fc80178b2cdf9e7795198e38f81dfff5002f70f06a" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.677618 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab155770-4f5d-43cf-a234-7c9394177ef6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.677656 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrg9h\" (UniqueName: \"kubernetes.io/projected/ab155770-4f5d-43cf-a234-7c9394177ef6-kube-api-access-vrg9h\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.698571 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab155770-4f5d-43cf-a234-7c9394177ef6-config-data" (OuterVolumeSpecName: "config-data") pod "ab155770-4f5d-43cf-a234-7c9394177ef6" (UID: "ab155770-4f5d-43cf-a234-7c9394177ef6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.714710 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.723234 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:04:07 crc kubenswrapper[4771]: E1002 10:04:07.723927 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8660c7e8-08fa-49fc-ab03-f0346caedce4" containerName="nova-metadata-metadata" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.723976 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8660c7e8-08fa-49fc-ab03-f0346caedce4" containerName="nova-metadata-metadata" Oct 02 10:04:07 crc kubenswrapper[4771]: E1002 10:04:07.723990 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab155770-4f5d-43cf-a234-7c9394177ef6" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.724000 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab155770-4f5d-43cf-a234-7c9394177ef6" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 10:04:07 crc kubenswrapper[4771]: E1002 10:04:07.724048 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8660c7e8-08fa-49fc-ab03-f0346caedce4" containerName="nova-metadata-log" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.724060 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8660c7e8-08fa-49fc-ab03-f0346caedce4" containerName="nova-metadata-log" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.724403 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8660c7e8-08fa-49fc-ab03-f0346caedce4" containerName="nova-metadata-metadata" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.724436 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8660c7e8-08fa-49fc-ab03-f0346caedce4" containerName="nova-metadata-log" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.724449 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab155770-4f5d-43cf-a234-7c9394177ef6" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.726005 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.733149 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.733863 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.754995 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.780510 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab155770-4f5d-43cf-a234-7c9394177ef6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.825503 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.846195 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.873372 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.875522 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.883843 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.883940 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn7pk\" (UniqueName: \"kubernetes.io/projected/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-kube-api-access-hn7pk\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.883969 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-logs\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.884094 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-config-data\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.884122 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.887288 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.892767 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.893014 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.893557 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.997266 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c49be1-946f-42eb-9767-6bb44807fb7d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.997466 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4wdg\" (UniqueName: \"kubernetes.io/projected/c8c49be1-946f-42eb-9767-6bb44807fb7d-kube-api-access-s4wdg\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.997536 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c49be1-946f-42eb-9767-6bb44807fb7d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.997743 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-config-data\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.997798 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.997900 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c49be1-946f-42eb-9767-6bb44807fb7d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.997950 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.997997 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c49be1-946f-42eb-9767-6bb44807fb7d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.998180 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn7pk\" (UniqueName: \"kubernetes.io/projected/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-kube-api-access-hn7pk\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.998230 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-logs\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:07 crc kubenswrapper[4771]: I1002 10:04:07.999851 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-logs\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.009858 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.012796 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-config-data\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.013572 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.061959 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn7pk\" (UniqueName: \"kubernetes.io/projected/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-kube-api-access-hn7pk\") pod \"nova-metadata-0\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " pod="openstack/nova-metadata-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.071959 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.100736 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c49be1-946f-42eb-9767-6bb44807fb7d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.100819 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c49be1-946f-42eb-9767-6bb44807fb7d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.100942 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c49be1-946f-42eb-9767-6bb44807fb7d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.101006 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4wdg\" (UniqueName: \"kubernetes.io/projected/c8c49be1-946f-42eb-9767-6bb44807fb7d-kube-api-access-s4wdg\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.101040 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c49be1-946f-42eb-9767-6bb44807fb7d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.105811 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c49be1-946f-42eb-9767-6bb44807fb7d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.119882 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c49be1-946f-42eb-9767-6bb44807fb7d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.122476 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c49be1-946f-42eb-9767-6bb44807fb7d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.127884 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c49be1-946f-42eb-9767-6bb44807fb7d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.173408 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4wdg\" (UniqueName: \"kubernetes.io/projected/c8c49be1-946f-42eb-9767-6bb44807fb7d-kube-api-access-s4wdg\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c49be1-946f-42eb-9767-6bb44807fb7d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.221976 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.538353 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8kv5" event={"ID":"8072221f-4513-4f85-b4ea-f06d0b81214c","Type":"ContainerStarted","Data":"3c4ab7901ba03aceea318c16b9ead4d48a7efd8780cde1a9045e9756819aa59f"} Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.544551 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc","Type":"ContainerStarted","Data":"7f6fbfc15f008d4c36872c95ba2428af7688dd4f5bf413f1417113f5a1c98e0f"} Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.594608 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q8kv5" podStartSLOduration=4.78352401 podStartE2EDuration="8.59458178s" podCreationTimestamp="2025-10-02 10:04:00 +0000 UTC" firstStartedPulling="2025-10-02 10:04:03.327117228 +0000 UTC m=+1630.974802295" lastFinishedPulling="2025-10-02 10:04:07.138174998 +0000 UTC m=+1634.785860065" observedRunningTime="2025-10-02 10:04:08.571655018 +0000 UTC m=+1636.219340095" watchObservedRunningTime="2025-10-02 10:04:08.59458178 +0000 UTC m=+1636.242266847" Oct 02 10:04:08 crc kubenswrapper[4771]: I1002 10:04:08.835369 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:04:09 crc kubenswrapper[4771]: I1002 10:04:09.045994 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:04:09 crc kubenswrapper[4771]: I1002 10:04:09.568873 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c8c49be1-946f-42eb-9767-6bb44807fb7d","Type":"ContainerStarted","Data":"6828daecaf4cf369b11e6a6191350c21fe9d2cb843aaf1e97eb98e6e99dbb565"} Oct 02 10:04:09 crc kubenswrapper[4771]: I1002 10:04:09.569403 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c8c49be1-946f-42eb-9767-6bb44807fb7d","Type":"ContainerStarted","Data":"9ca0d72d9bfc2705c749fa4307ad550ff98d91cdba22bead1a3babaf320637c1"} Oct 02 10:04:09 crc kubenswrapper[4771]: I1002 10:04:09.576677 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a17c7e4c-40e5-491f-bd4d-6e8a224c9421","Type":"ContainerStarted","Data":"0baa500f430ba2d9d5bebdb0339e4603a2dee13d24b5c04f29f73a1e780fad07"} Oct 02 10:04:09 crc kubenswrapper[4771]: I1002 10:04:09.576737 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a17c7e4c-40e5-491f-bd4d-6e8a224c9421","Type":"ContainerStarted","Data":"37db0e096ef5b1d568ed327b4432ca8ad5c9ced517a34620f11b77bb15b62835"} Oct 02 10:04:09 crc kubenswrapper[4771]: I1002 10:04:09.581308 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc","Type":"ContainerStarted","Data":"153c2127da2cae292d5bd4b9c6c7e26eab0676376ad8f0818fd8b1c31dac3cfa"} Oct 02 10:04:09 crc kubenswrapper[4771]: I1002 10:04:09.599032 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.5990114269999998 podStartE2EDuration="2.599011427s" podCreationTimestamp="2025-10-02 10:04:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:04:09.588809227 +0000 UTC m=+1637.236494294" watchObservedRunningTime="2025-10-02 10:04:09.599011427 +0000 UTC m=+1637.246696494" Oct 02 10:04:09 crc kubenswrapper[4771]: I1002 10:04:09.699690 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8660c7e8-08fa-49fc-ab03-f0346caedce4" path="/var/lib/kubelet/pods/8660c7e8-08fa-49fc-ab03-f0346caedce4/volumes" Oct 02 10:04:09 crc kubenswrapper[4771]: I1002 10:04:09.700352 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab155770-4f5d-43cf-a234-7c9394177ef6" path="/var/lib/kubelet/pods/ab155770-4f5d-43cf-a234-7c9394177ef6/volumes" Oct 02 10:04:10 crc kubenswrapper[4771]: I1002 10:04:10.595622 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a17c7e4c-40e5-491f-bd4d-6e8a224c9421","Type":"ContainerStarted","Data":"61d47adbfb336e1f32080454ef121b0ee51109ce466b0408832e451ee6ad66a8"} Oct 02 10:04:10 crc kubenswrapper[4771]: I1002 10:04:10.629638 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.629615018 podStartE2EDuration="3.629615018s" podCreationTimestamp="2025-10-02 10:04:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:04:10.615999681 +0000 UTC m=+1638.263684748" watchObservedRunningTime="2025-10-02 10:04:10.629615018 +0000 UTC m=+1638.277300085" Oct 02 10:04:10 crc kubenswrapper[4771]: I1002 10:04:10.910445 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:10 crc kubenswrapper[4771]: I1002 10:04:10.910752 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:10 crc kubenswrapper[4771]: I1002 10:04:10.981828 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:11 crc kubenswrapper[4771]: I1002 10:04:11.616647 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc","Type":"ContainerStarted","Data":"fb1f20c4441437bc04c904478af7a5c8637101ec7be58695bf54b56e0da77905"} Oct 02 10:04:11 crc kubenswrapper[4771]: I1002 10:04:11.617476 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="ceilometer-central-agent" containerID="cri-o://3c1f6add4476e5f41d38ed17aa09b9cff2d9afe4a31f03a8086de9fdbf88837f" gracePeriod=30 Oct 02 10:04:11 crc kubenswrapper[4771]: I1002 10:04:11.617518 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="proxy-httpd" containerID="cri-o://fb1f20c4441437bc04c904478af7a5c8637101ec7be58695bf54b56e0da77905" gracePeriod=30 Oct 02 10:04:11 crc kubenswrapper[4771]: I1002 10:04:11.617591 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="ceilometer-notification-agent" containerID="cri-o://7f6fbfc15f008d4c36872c95ba2428af7688dd4f5bf413f1417113f5a1c98e0f" gracePeriod=30 Oct 02 10:04:11 crc kubenswrapper[4771]: I1002 10:04:11.617613 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="sg-core" containerID="cri-o://153c2127da2cae292d5bd4b9c6c7e26eab0676376ad8f0818fd8b1c31dac3cfa" gracePeriod=30 Oct 02 10:04:11 crc kubenswrapper[4771]: I1002 10:04:11.654920 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.035329903 podStartE2EDuration="8.654898598s" podCreationTimestamp="2025-10-02 10:04:03 +0000 UTC" firstStartedPulling="2025-10-02 10:04:05.124296802 +0000 UTC m=+1632.771981869" lastFinishedPulling="2025-10-02 10:04:10.743865497 +0000 UTC m=+1638.391550564" observedRunningTime="2025-10-02 10:04:11.653023325 +0000 UTC m=+1639.300708392" watchObservedRunningTime="2025-10-02 10:04:11.654898598 +0000 UTC m=+1639.302583665" Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.254836 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.256106 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.256482 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.258388 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.630234 4771 generic.go:334] "Generic (PLEG): container finished" podID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerID="153c2127da2cae292d5bd4b9c6c7e26eab0676376ad8f0818fd8b1c31dac3cfa" exitCode=2 Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.630950 4771 generic.go:334] "Generic (PLEG): container finished" podID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerID="7f6fbfc15f008d4c36872c95ba2428af7688dd4f5bf413f1417113f5a1c98e0f" exitCode=0 Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.630306 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc","Type":"ContainerDied","Data":"153c2127da2cae292d5bd4b9c6c7e26eab0676376ad8f0818fd8b1c31dac3cfa"} Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.631107 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc","Type":"ContainerDied","Data":"7f6fbfc15f008d4c36872c95ba2428af7688dd4f5bf413f1417113f5a1c98e0f"} Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.631494 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.634671 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.855190 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-rzgws"] Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.862411 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.882810 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-rzgws"] Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.976591 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt5xt\" (UniqueName: \"kubernetes.io/projected/520c351c-db61-4f7b-8c3f-3239b619f553-kube-api-access-gt5xt\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.979499 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.979690 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.979889 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.980247 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-config\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:12 crc kubenswrapper[4771]: I1002 10:04:12.980383 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.072725 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.073065 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.082196 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.082254 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.082307 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.082346 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-config\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.082374 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.082428 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt5xt\" (UniqueName: \"kubernetes.io/projected/520c351c-db61-4f7b-8c3f-3239b619f553-kube-api-access-gt5xt\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.083998 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.084010 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-config\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.084188 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.084258 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.084695 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.104978 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt5xt\" (UniqueName: \"kubernetes.io/projected/520c351c-db61-4f7b-8c3f-3239b619f553-kube-api-access-gt5xt\") pod \"dnsmasq-dns-6b7bbf7cf9-rzgws\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.186816 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.223600 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.704994 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:04:13 crc kubenswrapper[4771]: E1002 10:04:13.706322 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:04:13 crc kubenswrapper[4771]: I1002 10:04:13.787993 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-rzgws"] Oct 02 10:04:14 crc kubenswrapper[4771]: I1002 10:04:14.662054 4771 generic.go:334] "Generic (PLEG): container finished" podID="520c351c-db61-4f7b-8c3f-3239b619f553" containerID="652826d725efd479f42346177f86c5e5979d27f61ea69ff16ef1062e8214140d" exitCode=0 Oct 02 10:04:14 crc kubenswrapper[4771]: I1002 10:04:14.662163 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" event={"ID":"520c351c-db61-4f7b-8c3f-3239b619f553","Type":"ContainerDied","Data":"652826d725efd479f42346177f86c5e5979d27f61ea69ff16ef1062e8214140d"} Oct 02 10:04:14 crc kubenswrapper[4771]: I1002 10:04:14.662542 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" event={"ID":"520c351c-db61-4f7b-8c3f-3239b619f553","Type":"ContainerStarted","Data":"7c22ad2044a31aaf13264b56d941c30d5a638f6ae0a8352ee37dc64641055cad"} Oct 02 10:04:15 crc kubenswrapper[4771]: I1002 10:04:15.713261 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:15 crc kubenswrapper[4771]: I1002 10:04:15.713631 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" event={"ID":"520c351c-db61-4f7b-8c3f-3239b619f553","Type":"ContainerStarted","Data":"c6e483c524ae041bc3f5309f41b8d5f92d6d95332ef9a6be6377b18fc73abd3d"} Oct 02 10:04:15 crc kubenswrapper[4771]: I1002 10:04:15.737975 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" podStartSLOduration=3.737951357 podStartE2EDuration="3.737951357s" podCreationTimestamp="2025-10-02 10:04:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:04:15.722982131 +0000 UTC m=+1643.370667198" watchObservedRunningTime="2025-10-02 10:04:15.737951357 +0000 UTC m=+1643.385636424" Oct 02 10:04:16 crc kubenswrapper[4771]: I1002 10:04:16.017794 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:04:16 crc kubenswrapper[4771]: I1002 10:04:16.018530 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="78504a04-3786-414a-94f9-e7cd45d2deee" containerName="nova-api-api" containerID="cri-o://ad62c8eb290c27c2af0e52f0b674e9aee22b87f31fbab7a32dfdf97e467f834e" gracePeriod=30 Oct 02 10:04:16 crc kubenswrapper[4771]: I1002 10:04:16.018452 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="78504a04-3786-414a-94f9-e7cd45d2deee" containerName="nova-api-log" containerID="cri-o://91f1ccf5a30da789ef16da87b41d107f570e92baafa4e58ff49db615a66e6e1d" gracePeriod=30 Oct 02 10:04:16 crc kubenswrapper[4771]: E1002 10:04:16.338185 4771 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/5babe04d3d26cc47fc6b896d0ca2ddd825b20b99b3116dcb4d5fcfcceef7bc41/diff" to get inode usage: stat /var/lib/containers/storage/overlay/5babe04d3d26cc47fc6b896d0ca2ddd825b20b99b3116dcb4d5fcfcceef7bc41/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_9158fae1-e129-47d9-af54-2105c7d79e41/ceilometer-central-agent/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_9158fae1-e129-47d9-af54-2105c7d79e41/ceilometer-central-agent/0.log: no such file or directory Oct 02 10:04:16 crc kubenswrapper[4771]: I1002 10:04:16.717564 4771 generic.go:334] "Generic (PLEG): container finished" podID="78504a04-3786-414a-94f9-e7cd45d2deee" containerID="91f1ccf5a30da789ef16da87b41d107f570e92baafa4e58ff49db615a66e6e1d" exitCode=143 Oct 02 10:04:16 crc kubenswrapper[4771]: I1002 10:04:16.717668 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78504a04-3786-414a-94f9-e7cd45d2deee","Type":"ContainerDied","Data":"91f1ccf5a30da789ef16da87b41d107f570e92baafa4e58ff49db615a66e6e1d"} Oct 02 10:04:17 crc kubenswrapper[4771]: E1002 10:04:17.118580 4771 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/6463256ce16aaaeb06a80e648ed1bdb1bb7d6044ee87b06c1f89a9b75c094262/diff" to get inode usage: stat /var/lib/containers/storage/overlay/6463256ce16aaaeb06a80e648ed1bdb1bb7d6044ee87b06c1f89a9b75c094262/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_9158fae1-e129-47d9-af54-2105c7d79e41/ceilometer-notification-agent/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_9158fae1-e129-47d9-af54-2105c7d79e41/ceilometer-notification-agent/0.log: no such file or directory Oct 02 10:04:17 crc kubenswrapper[4771]: I1002 10:04:17.748183 4771 generic.go:334] "Generic (PLEG): container finished" podID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerID="3c1f6add4476e5f41d38ed17aa09b9cff2d9afe4a31f03a8086de9fdbf88837f" exitCode=0 Oct 02 10:04:17 crc kubenswrapper[4771]: I1002 10:04:17.748241 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc","Type":"ContainerDied","Data":"3c1f6add4476e5f41d38ed17aa09b9cff2d9afe4a31f03a8086de9fdbf88837f"} Oct 02 10:04:17 crc kubenswrapper[4771]: E1002 10:04:17.957958 4771 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/cc0176e3dc6a47e852967cb5ea4b82d143f8ebc485daa1e581fd6f6b3d8f1aba/diff" to get inode usage: stat /var/lib/containers/storage/overlay/cc0176e3dc6a47e852967cb5ea4b82d143f8ebc485daa1e581fd6f6b3d8f1aba/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_9158fae1-e129-47d9-af54-2105c7d79e41/sg-core/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_9158fae1-e129-47d9-af54-2105c7d79e41/sg-core/0.log: no such file or directory Oct 02 10:04:18 crc kubenswrapper[4771]: I1002 10:04:18.073097 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 10:04:18 crc kubenswrapper[4771]: I1002 10:04:18.073455 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 10:04:18 crc kubenswrapper[4771]: I1002 10:04:18.223168 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:18 crc kubenswrapper[4771]: I1002 10:04:18.264318 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:18 crc kubenswrapper[4771]: I1002 10:04:18.779391 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:04:18 crc kubenswrapper[4771]: I1002 10:04:18.985030 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-96p7k"] Oct 02 10:04:18 crc kubenswrapper[4771]: I1002 10:04:18.987531 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:18 crc kubenswrapper[4771]: I1002 10:04:18.996413 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.000499 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-96p7k"] Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.000992 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.085932 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4vg5\" (UniqueName: \"kubernetes.io/projected/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-kube-api-access-m4vg5\") pod \"nova-cell1-cell-mapping-96p7k\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.086006 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-scripts\") pod \"nova-cell1-cell-mapping-96p7k\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.086244 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-96p7k\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.086283 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-config-data\") pod \"nova-cell1-cell-mapping-96p7k\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.088428 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.254:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.088463 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.254:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.188925 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4vg5\" (UniqueName: \"kubernetes.io/projected/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-kube-api-access-m4vg5\") pod \"nova-cell1-cell-mapping-96p7k\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.189242 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-scripts\") pod \"nova-cell1-cell-mapping-96p7k\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.189469 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-96p7k\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.189582 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-config-data\") pod \"nova-cell1-cell-mapping-96p7k\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.196517 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-scripts\") pod \"nova-cell1-cell-mapping-96p7k\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.200855 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-config-data\") pod \"nova-cell1-cell-mapping-96p7k\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.201246 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-96p7k\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.208980 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4vg5\" (UniqueName: \"kubernetes.io/projected/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-kube-api-access-m4vg5\") pod \"nova-cell1-cell-mapping-96p7k\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.320059 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:19 crc kubenswrapper[4771]: E1002 10:04:19.327846 4771 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/7d330b11d6c8a321b9c3a0c3dafb1e3b8ef4e89ff36bebb786c76aa0d9c0e2ad/diff" to get inode usage: stat /var/lib/containers/storage/overlay/7d330b11d6c8a321b9c3a0c3dafb1e3b8ef4e89ff36bebb786c76aa0d9c0e2ad/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_9158fae1-e129-47d9-af54-2105c7d79e41/proxy-httpd/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_9158fae1-e129-47d9-af54-2105c7d79e41/proxy-httpd/0.log: no such file or directory Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.801449 4771 generic.go:334] "Generic (PLEG): container finished" podID="78504a04-3786-414a-94f9-e7cd45d2deee" containerID="ad62c8eb290c27c2af0e52f0b674e9aee22b87f31fbab7a32dfdf97e467f834e" exitCode=0 Oct 02 10:04:19 crc kubenswrapper[4771]: I1002 10:04:19.802651 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78504a04-3786-414a-94f9-e7cd45d2deee","Type":"ContainerDied","Data":"ad62c8eb290c27c2af0e52f0b674e9aee22b87f31fbab7a32dfdf97e467f834e"} Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.246752 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-96p7k"] Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.412916 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.534504 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78504a04-3786-414a-94f9-e7cd45d2deee-config-data\") pod \"78504a04-3786-414a-94f9-e7cd45d2deee\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.534749 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78504a04-3786-414a-94f9-e7cd45d2deee-combined-ca-bundle\") pod \"78504a04-3786-414a-94f9-e7cd45d2deee\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.534837 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78504a04-3786-414a-94f9-e7cd45d2deee-logs\") pod \"78504a04-3786-414a-94f9-e7cd45d2deee\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.534993 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxxnd\" (UniqueName: \"kubernetes.io/projected/78504a04-3786-414a-94f9-e7cd45d2deee-kube-api-access-wxxnd\") pod \"78504a04-3786-414a-94f9-e7cd45d2deee\" (UID: \"78504a04-3786-414a-94f9-e7cd45d2deee\") " Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.536808 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78504a04-3786-414a-94f9-e7cd45d2deee-logs" (OuterVolumeSpecName: "logs") pod "78504a04-3786-414a-94f9-e7cd45d2deee" (UID: "78504a04-3786-414a-94f9-e7cd45d2deee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.564547 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78504a04-3786-414a-94f9-e7cd45d2deee-kube-api-access-wxxnd" (OuterVolumeSpecName: "kube-api-access-wxxnd") pod "78504a04-3786-414a-94f9-e7cd45d2deee" (UID: "78504a04-3786-414a-94f9-e7cd45d2deee"). InnerVolumeSpecName "kube-api-access-wxxnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.605247 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78504a04-3786-414a-94f9-e7cd45d2deee-config-data" (OuterVolumeSpecName: "config-data") pod "78504a04-3786-414a-94f9-e7cd45d2deee" (UID: "78504a04-3786-414a-94f9-e7cd45d2deee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.605792 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78504a04-3786-414a-94f9-e7cd45d2deee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78504a04-3786-414a-94f9-e7cd45d2deee" (UID: "78504a04-3786-414a-94f9-e7cd45d2deee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.638433 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78504a04-3786-414a-94f9-e7cd45d2deee-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.638473 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78504a04-3786-414a-94f9-e7cd45d2deee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.638488 4771 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78504a04-3786-414a-94f9-e7cd45d2deee-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.638497 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxxnd\" (UniqueName: \"kubernetes.io/projected/78504a04-3786-414a-94f9-e7cd45d2deee-kube-api-access-wxxnd\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.852504 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78504a04-3786-414a-94f9-e7cd45d2deee","Type":"ContainerDied","Data":"807be8f229569cbc1cd7fff6b0f5458a1633c2b343c1e1b0310d6757c31557f8"} Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.852888 4771 scope.go:117] "RemoveContainer" containerID="ad62c8eb290c27c2af0e52f0b674e9aee22b87f31fbab7a32dfdf97e467f834e" Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.853090 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.857445 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-96p7k" event={"ID":"8da47e22-e6c2-4c9f-92f0-9ae49a82be69","Type":"ContainerStarted","Data":"1756439a79a732de29371d863cdf56e05a0e7a501791090ddfd222d20fe54f55"} Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.917821 4771 scope.go:117] "RemoveContainer" containerID="91f1ccf5a30da789ef16da87b41d107f570e92baafa4e58ff49db615a66e6e1d" Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.957575 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:04:20 crc kubenswrapper[4771]: I1002 10:04:20.981736 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:20.998813 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 10:04:21 crc kubenswrapper[4771]: E1002 10:04:20.999491 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78504a04-3786-414a-94f9-e7cd45d2deee" containerName="nova-api-api" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:20.999509 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="78504a04-3786-414a-94f9-e7cd45d2deee" containerName="nova-api-api" Oct 02 10:04:21 crc kubenswrapper[4771]: E1002 10:04:20.999520 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78504a04-3786-414a-94f9-e7cd45d2deee" containerName="nova-api-log" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:20.999527 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="78504a04-3786-414a-94f9-e7cd45d2deee" containerName="nova-api-log" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:20.999951 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="78504a04-3786-414a-94f9-e7cd45d2deee" containerName="nova-api-api" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:20.999996 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="78504a04-3786-414a-94f9-e7cd45d2deee" containerName="nova-api-log" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.001684 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.010029 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.010545 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.010723 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.053298 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.053446 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-public-tls-certs\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.053464 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gdsk\" (UniqueName: \"kubernetes.io/projected/f1f84431-dda1-43c4-8246-5742c470bf09-kube-api-access-5gdsk\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.053546 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.053564 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-config-data\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.053593 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1f84431-dda1-43c4-8246-5742c470bf09-logs\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.080617 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.118411 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.156034 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1f84431-dda1-43c4-8246-5742c470bf09-logs\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.156480 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.156515 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1f84431-dda1-43c4-8246-5742c470bf09-logs\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.156671 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-public-tls-certs\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.156693 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gdsk\" (UniqueName: \"kubernetes.io/projected/f1f84431-dda1-43c4-8246-5742c470bf09-kube-api-access-5gdsk\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.156819 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.156836 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-config-data\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.161692 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.163408 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-config-data\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.164870 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.165792 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-public-tls-certs\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.184560 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gdsk\" (UniqueName: \"kubernetes.io/projected/f1f84431-dda1-43c4-8246-5742c470bf09-kube-api-access-5gdsk\") pod \"nova-api-0\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.192815 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8kv5"] Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.388425 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.707597 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78504a04-3786-414a-94f9-e7cd45d2deee" path="/var/lib/kubelet/pods/78504a04-3786-414a-94f9-e7cd45d2deee/volumes" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.873393 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-96p7k" event={"ID":"8da47e22-e6c2-4c9f-92f0-9ae49a82be69","Type":"ContainerStarted","Data":"0183f6869b55796a23e6c706865698dad8d57f152055b84e7b6a0e7d138391c6"} Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.874834 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q8kv5" podUID="8072221f-4513-4f85-b4ea-f06d0b81214c" containerName="registry-server" containerID="cri-o://3c4ab7901ba03aceea318c16b9ead4d48a7efd8780cde1a9045e9756819aa59f" gracePeriod=2 Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.893444 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-96p7k" podStartSLOduration=3.893427704 podStartE2EDuration="3.893427704s" podCreationTimestamp="2025-10-02 10:04:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:04:21.888812315 +0000 UTC m=+1649.536497372" watchObservedRunningTime="2025-10-02 10:04:21.893427704 +0000 UTC m=+1649.541112771" Oct 02 10:04:21 crc kubenswrapper[4771]: I1002 10:04:21.953943 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.836956 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.889488 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1f84431-dda1-43c4-8246-5742c470bf09","Type":"ContainerStarted","Data":"2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8"} Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.889552 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1f84431-dda1-43c4-8246-5742c470bf09","Type":"ContainerStarted","Data":"93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7"} Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.889567 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1f84431-dda1-43c4-8246-5742c470bf09","Type":"ContainerStarted","Data":"1b44444531bad32b55b717a002455981e055ec328e84e6121c4b2a5fc6908d6e"} Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.899413 4771 generic.go:334] "Generic (PLEG): container finished" podID="8072221f-4513-4f85-b4ea-f06d0b81214c" containerID="3c4ab7901ba03aceea318c16b9ead4d48a7efd8780cde1a9045e9756819aa59f" exitCode=0 Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.900505 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8kv5" Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.900736 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8kv5" event={"ID":"8072221f-4513-4f85-b4ea-f06d0b81214c","Type":"ContainerDied","Data":"3c4ab7901ba03aceea318c16b9ead4d48a7efd8780cde1a9045e9756819aa59f"} Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.900771 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8kv5" event={"ID":"8072221f-4513-4f85-b4ea-f06d0b81214c","Type":"ContainerDied","Data":"d5af6363beb94a27d4291c0997172881647c08e567470c80e59369ce5b714b99"} Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.900788 4771 scope.go:117] "RemoveContainer" containerID="3c4ab7901ba03aceea318c16b9ead4d48a7efd8780cde1a9045e9756819aa59f" Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.916942 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.916916543 podStartE2EDuration="2.916916543s" podCreationTimestamp="2025-10-02 10:04:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:04:22.911361589 +0000 UTC m=+1650.559046686" watchObservedRunningTime="2025-10-02 10:04:22.916916543 +0000 UTC m=+1650.564601620" Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.918019 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lt6c\" (UniqueName: \"kubernetes.io/projected/8072221f-4513-4f85-b4ea-f06d0b81214c-kube-api-access-4lt6c\") pod \"8072221f-4513-4f85-b4ea-f06d0b81214c\" (UID: \"8072221f-4513-4f85-b4ea-f06d0b81214c\") " Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.918102 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8072221f-4513-4f85-b4ea-f06d0b81214c-catalog-content\") pod \"8072221f-4513-4f85-b4ea-f06d0b81214c\" (UID: \"8072221f-4513-4f85-b4ea-f06d0b81214c\") " Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.918342 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8072221f-4513-4f85-b4ea-f06d0b81214c-utilities\") pod \"8072221f-4513-4f85-b4ea-f06d0b81214c\" (UID: \"8072221f-4513-4f85-b4ea-f06d0b81214c\") " Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.919828 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8072221f-4513-4f85-b4ea-f06d0b81214c-utilities" (OuterVolumeSpecName: "utilities") pod "8072221f-4513-4f85-b4ea-f06d0b81214c" (UID: "8072221f-4513-4f85-b4ea-f06d0b81214c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.925405 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8072221f-4513-4f85-b4ea-f06d0b81214c-kube-api-access-4lt6c" (OuterVolumeSpecName: "kube-api-access-4lt6c") pod "8072221f-4513-4f85-b4ea-f06d0b81214c" (UID: "8072221f-4513-4f85-b4ea-f06d0b81214c"). InnerVolumeSpecName "kube-api-access-4lt6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.941655 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8072221f-4513-4f85-b4ea-f06d0b81214c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8072221f-4513-4f85-b4ea-f06d0b81214c" (UID: "8072221f-4513-4f85-b4ea-f06d0b81214c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:04:22 crc kubenswrapper[4771]: I1002 10:04:22.981670 4771 scope.go:117] "RemoveContainer" containerID="e08dea8c94e8f50f4c57abd2be2bb92804b75ef06bf9fcff15d36c817675f59b" Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.022399 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8072221f-4513-4f85-b4ea-f06d0b81214c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.022445 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lt6c\" (UniqueName: \"kubernetes.io/projected/8072221f-4513-4f85-b4ea-f06d0b81214c-kube-api-access-4lt6c\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.022461 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8072221f-4513-4f85-b4ea-f06d0b81214c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.064742 4771 scope.go:117] "RemoveContainer" containerID="b643df3d0efcea91de0fc7906e4b194ded0b5e1777bf0eaf10e4026496870509" Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.126501 4771 scope.go:117] "RemoveContainer" containerID="3c4ab7901ba03aceea318c16b9ead4d48a7efd8780cde1a9045e9756819aa59f" Oct 02 10:04:23 crc kubenswrapper[4771]: E1002 10:04:23.127310 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c4ab7901ba03aceea318c16b9ead4d48a7efd8780cde1a9045e9756819aa59f\": container with ID starting with 3c4ab7901ba03aceea318c16b9ead4d48a7efd8780cde1a9045e9756819aa59f not found: ID does not exist" containerID="3c4ab7901ba03aceea318c16b9ead4d48a7efd8780cde1a9045e9756819aa59f" Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.127346 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c4ab7901ba03aceea318c16b9ead4d48a7efd8780cde1a9045e9756819aa59f"} err="failed to get container status \"3c4ab7901ba03aceea318c16b9ead4d48a7efd8780cde1a9045e9756819aa59f\": rpc error: code = NotFound desc = could not find container \"3c4ab7901ba03aceea318c16b9ead4d48a7efd8780cde1a9045e9756819aa59f\": container with ID starting with 3c4ab7901ba03aceea318c16b9ead4d48a7efd8780cde1a9045e9756819aa59f not found: ID does not exist" Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.127370 4771 scope.go:117] "RemoveContainer" containerID="e08dea8c94e8f50f4c57abd2be2bb92804b75ef06bf9fcff15d36c817675f59b" Oct 02 10:04:23 crc kubenswrapper[4771]: E1002 10:04:23.127933 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e08dea8c94e8f50f4c57abd2be2bb92804b75ef06bf9fcff15d36c817675f59b\": container with ID starting with e08dea8c94e8f50f4c57abd2be2bb92804b75ef06bf9fcff15d36c817675f59b not found: ID does not exist" containerID="e08dea8c94e8f50f4c57abd2be2bb92804b75ef06bf9fcff15d36c817675f59b" Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.127966 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e08dea8c94e8f50f4c57abd2be2bb92804b75ef06bf9fcff15d36c817675f59b"} err="failed to get container status \"e08dea8c94e8f50f4c57abd2be2bb92804b75ef06bf9fcff15d36c817675f59b\": rpc error: code = NotFound desc = could not find container \"e08dea8c94e8f50f4c57abd2be2bb92804b75ef06bf9fcff15d36c817675f59b\": container with ID starting with e08dea8c94e8f50f4c57abd2be2bb92804b75ef06bf9fcff15d36c817675f59b not found: ID does not exist" Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.128000 4771 scope.go:117] "RemoveContainer" containerID="b643df3d0efcea91de0fc7906e4b194ded0b5e1777bf0eaf10e4026496870509" Oct 02 10:04:23 crc kubenswrapper[4771]: E1002 10:04:23.128584 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b643df3d0efcea91de0fc7906e4b194ded0b5e1777bf0eaf10e4026496870509\": container with ID starting with b643df3d0efcea91de0fc7906e4b194ded0b5e1777bf0eaf10e4026496870509 not found: ID does not exist" containerID="b643df3d0efcea91de0fc7906e4b194ded0b5e1777bf0eaf10e4026496870509" Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.128747 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b643df3d0efcea91de0fc7906e4b194ded0b5e1777bf0eaf10e4026496870509"} err="failed to get container status \"b643df3d0efcea91de0fc7906e4b194ded0b5e1777bf0eaf10e4026496870509\": rpc error: code = NotFound desc = could not find container \"b643df3d0efcea91de0fc7906e4b194ded0b5e1777bf0eaf10e4026496870509\": container with ID starting with b643df3d0efcea91de0fc7906e4b194ded0b5e1777bf0eaf10e4026496870509 not found: ID does not exist" Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.188479 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.312519 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-gphqv"] Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.312774 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" podUID="7ccdd3d7-717c-4879-9155-41669900240b" containerName="dnsmasq-dns" containerID="cri-o://884149544bd01e9f83b4a17b66cd975a805ee6d96fc50f266ff9351ca1217d87" gracePeriod=10 Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.328985 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8kv5"] Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.350165 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8kv5"] Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.701634 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8072221f-4513-4f85-b4ea-f06d0b81214c" path="/var/lib/kubelet/pods/8072221f-4513-4f85-b4ea-f06d0b81214c/volumes" Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.930489 4771 generic.go:334] "Generic (PLEG): container finished" podID="7ccdd3d7-717c-4879-9155-41669900240b" containerID="884149544bd01e9f83b4a17b66cd975a805ee6d96fc50f266ff9351ca1217d87" exitCode=0 Oct 02 10:04:23 crc kubenswrapper[4771]: I1002 10:04:23.930604 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" event={"ID":"7ccdd3d7-717c-4879-9155-41669900240b","Type":"ContainerDied","Data":"884149544bd01e9f83b4a17b66cd975a805ee6d96fc50f266ff9351ca1217d87"} Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.228534 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.369432 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-dns-svc\") pod \"7ccdd3d7-717c-4879-9155-41669900240b\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.369540 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4zph\" (UniqueName: \"kubernetes.io/projected/7ccdd3d7-717c-4879-9155-41669900240b-kube-api-access-f4zph\") pod \"7ccdd3d7-717c-4879-9155-41669900240b\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.369659 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-ovsdbserver-nb\") pod \"7ccdd3d7-717c-4879-9155-41669900240b\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.369718 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-dns-swift-storage-0\") pod \"7ccdd3d7-717c-4879-9155-41669900240b\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.369773 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-ovsdbserver-sb\") pod \"7ccdd3d7-717c-4879-9155-41669900240b\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.369824 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-config\") pod \"7ccdd3d7-717c-4879-9155-41669900240b\" (UID: \"7ccdd3d7-717c-4879-9155-41669900240b\") " Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.395667 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ccdd3d7-717c-4879-9155-41669900240b-kube-api-access-f4zph" (OuterVolumeSpecName: "kube-api-access-f4zph") pod "7ccdd3d7-717c-4879-9155-41669900240b" (UID: "7ccdd3d7-717c-4879-9155-41669900240b"). InnerVolumeSpecName "kube-api-access-f4zph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.446007 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7ccdd3d7-717c-4879-9155-41669900240b" (UID: "7ccdd3d7-717c-4879-9155-41669900240b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.470960 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7ccdd3d7-717c-4879-9155-41669900240b" (UID: "7ccdd3d7-717c-4879-9155-41669900240b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.472364 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-config" (OuterVolumeSpecName: "config") pod "7ccdd3d7-717c-4879-9155-41669900240b" (UID: "7ccdd3d7-717c-4879-9155-41669900240b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.474368 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.474392 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4zph\" (UniqueName: \"kubernetes.io/projected/7ccdd3d7-717c-4879-9155-41669900240b-kube-api-access-f4zph\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.474405 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.474472 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.479899 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7ccdd3d7-717c-4879-9155-41669900240b" (UID: "7ccdd3d7-717c-4879-9155-41669900240b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.498170 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7ccdd3d7-717c-4879-9155-41669900240b" (UID: "7ccdd3d7-717c-4879-9155-41669900240b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.576391 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.576425 4771 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ccdd3d7-717c-4879-9155-41669900240b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.985551 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" event={"ID":"7ccdd3d7-717c-4879-9155-41669900240b","Type":"ContainerDied","Data":"8be4f00e45f1892f0d13272e956df3718b990d43e014fa8f5cdf6023b849dd54"} Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.986014 4771 scope.go:117] "RemoveContainer" containerID="884149544bd01e9f83b4a17b66cd975a805ee6d96fc50f266ff9351ca1217d87" Oct 02 10:04:24 crc kubenswrapper[4771]: I1002 10:04:24.986330 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" Oct 02 10:04:25 crc kubenswrapper[4771]: I1002 10:04:25.032405 4771 scope.go:117] "RemoveContainer" containerID="b387454540528deec2342f4d808d6306bdb8825a835d5715c0efe51766e1b363" Oct 02 10:04:25 crc kubenswrapper[4771]: I1002 10:04:25.056185 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-gphqv"] Oct 02 10:04:25 crc kubenswrapper[4771]: I1002 10:04:25.072498 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-gphqv"] Oct 02 10:04:25 crc kubenswrapper[4771]: I1002 10:04:25.696689 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ccdd3d7-717c-4879-9155-41669900240b" path="/var/lib/kubelet/pods/7ccdd3d7-717c-4879-9155-41669900240b/volumes" Oct 02 10:04:27 crc kubenswrapper[4771]: I1002 10:04:27.681460 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:04:27 crc kubenswrapper[4771]: E1002 10:04:27.681989 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:04:28 crc kubenswrapper[4771]: I1002 10:04:28.049808 4771 generic.go:334] "Generic (PLEG): container finished" podID="8da47e22-e6c2-4c9f-92f0-9ae49a82be69" containerID="0183f6869b55796a23e6c706865698dad8d57f152055b84e7b6a0e7d138391c6" exitCode=0 Oct 02 10:04:28 crc kubenswrapper[4771]: I1002 10:04:28.049870 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-96p7k" event={"ID":"8da47e22-e6c2-4c9f-92f0-9ae49a82be69","Type":"ContainerDied","Data":"0183f6869b55796a23e6c706865698dad8d57f152055b84e7b6a0e7d138391c6"} Oct 02 10:04:28 crc kubenswrapper[4771]: I1002 10:04:28.077870 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 10:04:28 crc kubenswrapper[4771]: I1002 10:04:28.078355 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 10:04:28 crc kubenswrapper[4771]: I1002 10:04:28.088389 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 10:04:28 crc kubenswrapper[4771]: I1002 10:04:28.733956 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-9b86998b5-gphqv" podUID="7ccdd3d7-717c-4879-9155-41669900240b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.242:5353: i/o timeout" Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.094341 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.745235 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.809644 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-config-data\") pod \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.809757 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4vg5\" (UniqueName: \"kubernetes.io/projected/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-kube-api-access-m4vg5\") pod \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.809822 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-combined-ca-bundle\") pod \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.809852 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-scripts\") pod \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\" (UID: \"8da47e22-e6c2-4c9f-92f0-9ae49a82be69\") " Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.816457 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-scripts" (OuterVolumeSpecName: "scripts") pod "8da47e22-e6c2-4c9f-92f0-9ae49a82be69" (UID: "8da47e22-e6c2-4c9f-92f0-9ae49a82be69"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.817073 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-kube-api-access-m4vg5" (OuterVolumeSpecName: "kube-api-access-m4vg5") pod "8da47e22-e6c2-4c9f-92f0-9ae49a82be69" (UID: "8da47e22-e6c2-4c9f-92f0-9ae49a82be69"). InnerVolumeSpecName "kube-api-access-m4vg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.851911 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8da47e22-e6c2-4c9f-92f0-9ae49a82be69" (UID: "8da47e22-e6c2-4c9f-92f0-9ae49a82be69"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.854757 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-config-data" (OuterVolumeSpecName: "config-data") pod "8da47e22-e6c2-4c9f-92f0-9ae49a82be69" (UID: "8da47e22-e6c2-4c9f-92f0-9ae49a82be69"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.913165 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.913203 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4vg5\" (UniqueName: \"kubernetes.io/projected/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-kube-api-access-m4vg5\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.913213 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:29 crc kubenswrapper[4771]: I1002 10:04:29.913222 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8da47e22-e6c2-4c9f-92f0-9ae49a82be69-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:30 crc kubenswrapper[4771]: I1002 10:04:30.098141 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-96p7k" event={"ID":"8da47e22-e6c2-4c9f-92f0-9ae49a82be69","Type":"ContainerDied","Data":"1756439a79a732de29371d863cdf56e05a0e7a501791090ddfd222d20fe54f55"} Oct 02 10:04:30 crc kubenswrapper[4771]: I1002 10:04:30.098551 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1756439a79a732de29371d863cdf56e05a0e7a501791090ddfd222d20fe54f55" Oct 02 10:04:30 crc kubenswrapper[4771]: I1002 10:04:30.098183 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-96p7k" Oct 02 10:04:30 crc kubenswrapper[4771]: I1002 10:04:30.277303 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:04:30 crc kubenswrapper[4771]: I1002 10:04:30.277575 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f1f84431-dda1-43c4-8246-5742c470bf09" containerName="nova-api-log" containerID="cri-o://93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7" gracePeriod=30 Oct 02 10:04:30 crc kubenswrapper[4771]: I1002 10:04:30.277678 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f1f84431-dda1-43c4-8246-5742c470bf09" containerName="nova-api-api" containerID="cri-o://2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8" gracePeriod=30 Oct 02 10:04:30 crc kubenswrapper[4771]: I1002 10:04:30.304958 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:04:30 crc kubenswrapper[4771]: I1002 10:04:30.305344 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d343eb2f-faff-47b8-aad6-9d24a3fc39c7" containerName="nova-scheduler-scheduler" containerID="cri-o://eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911" gracePeriod=30 Oct 02 10:04:30 crc kubenswrapper[4771]: I1002 10:04:30.344996 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:04:30 crc kubenswrapper[4771]: E1002 10:04:30.538204 4771 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/6826be9d5e695829fba2a02002264d3067d4070d08ab2b0303ba265e1b2ab52f/diff" to get inode usage: stat /var/lib/containers/storage/overlay/6826be9d5e695829fba2a02002264d3067d4070d08ab2b0303ba265e1b2ab52f/diff: no such file or directory, extraDiskErr: Oct 02 10:04:30 crc kubenswrapper[4771]: E1002 10:04:30.914404 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:04:30 crc kubenswrapper[4771]: E1002 10:04:30.916910 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:04:30 crc kubenswrapper[4771]: E1002 10:04:30.918726 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:04:30 crc kubenswrapper[4771]: E1002 10:04:30.918796 4771 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="d343eb2f-faff-47b8-aad6-9d24a3fc39c7" containerName="nova-scheduler-scheduler" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.103597 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.111999 4771 generic.go:334] "Generic (PLEG): container finished" podID="f1f84431-dda1-43c4-8246-5742c470bf09" containerID="2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8" exitCode=0 Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.112036 4771 generic.go:334] "Generic (PLEG): container finished" podID="f1f84431-dda1-43c4-8246-5742c470bf09" containerID="93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7" exitCode=143 Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.112049 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1f84431-dda1-43c4-8246-5742c470bf09","Type":"ContainerDied","Data":"2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8"} Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.112092 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.112106 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1f84431-dda1-43c4-8246-5742c470bf09","Type":"ContainerDied","Data":"93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7"} Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.112118 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1f84431-dda1-43c4-8246-5742c470bf09","Type":"ContainerDied","Data":"1b44444531bad32b55b717a002455981e055ec328e84e6121c4b2a5fc6908d6e"} Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.112150 4771 scope.go:117] "RemoveContainer" containerID="2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.140296 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1f84431-dda1-43c4-8246-5742c470bf09-logs\") pod \"f1f84431-dda1-43c4-8246-5742c470bf09\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.140383 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-combined-ca-bundle\") pod \"f1f84431-dda1-43c4-8246-5742c470bf09\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.140439 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-public-tls-certs\") pod \"f1f84431-dda1-43c4-8246-5742c470bf09\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.140489 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-internal-tls-certs\") pod \"f1f84431-dda1-43c4-8246-5742c470bf09\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.140632 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gdsk\" (UniqueName: \"kubernetes.io/projected/f1f84431-dda1-43c4-8246-5742c470bf09-kube-api-access-5gdsk\") pod \"f1f84431-dda1-43c4-8246-5742c470bf09\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.140702 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1f84431-dda1-43c4-8246-5742c470bf09-logs" (OuterVolumeSpecName: "logs") pod "f1f84431-dda1-43c4-8246-5742c470bf09" (UID: "f1f84431-dda1-43c4-8246-5742c470bf09"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.140783 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-config-data\") pod \"f1f84431-dda1-43c4-8246-5742c470bf09\" (UID: \"f1f84431-dda1-43c4-8246-5742c470bf09\") " Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.141836 4771 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1f84431-dda1-43c4-8246-5742c470bf09-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.147938 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1f84431-dda1-43c4-8246-5742c470bf09-kube-api-access-5gdsk" (OuterVolumeSpecName: "kube-api-access-5gdsk") pod "f1f84431-dda1-43c4-8246-5742c470bf09" (UID: "f1f84431-dda1-43c4-8246-5742c470bf09"). InnerVolumeSpecName "kube-api-access-5gdsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.151016 4771 scope.go:117] "RemoveContainer" containerID="93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.247254 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gdsk\" (UniqueName: \"kubernetes.io/projected/f1f84431-dda1-43c4-8246-5742c470bf09-kube-api-access-5gdsk\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.247647 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1f84431-dda1-43c4-8246-5742c470bf09" (UID: "f1f84431-dda1-43c4-8246-5742c470bf09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.250392 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-config-data" (OuterVolumeSpecName: "config-data") pod "f1f84431-dda1-43c4-8246-5742c470bf09" (UID: "f1f84431-dda1-43c4-8246-5742c470bf09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.297875 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f1f84431-dda1-43c4-8246-5742c470bf09" (UID: "f1f84431-dda1-43c4-8246-5742c470bf09"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.306362 4771 scope.go:117] "RemoveContainer" containerID="2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8" Oct 02 10:04:31 crc kubenswrapper[4771]: E1002 10:04:31.308038 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8\": container with ID starting with 2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8 not found: ID does not exist" containerID="2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.308072 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8"} err="failed to get container status \"2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8\": rpc error: code = NotFound desc = could not find container \"2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8\": container with ID starting with 2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8 not found: ID does not exist" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.308114 4771 scope.go:117] "RemoveContainer" containerID="93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7" Oct 02 10:04:31 crc kubenswrapper[4771]: E1002 10:04:31.308501 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7\": container with ID starting with 93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7 not found: ID does not exist" containerID="93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.308571 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7"} err="failed to get container status \"93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7\": rpc error: code = NotFound desc = could not find container \"93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7\": container with ID starting with 93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7 not found: ID does not exist" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.308591 4771 scope.go:117] "RemoveContainer" containerID="2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.308852 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8"} err="failed to get container status \"2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8\": rpc error: code = NotFound desc = could not find container \"2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8\": container with ID starting with 2c20f0c63cdf38b945c793d76d643b7717e31abc7cc2e3757d7529a2bd6a99f8 not found: ID does not exist" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.308891 4771 scope.go:117] "RemoveContainer" containerID="93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.309146 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7"} err="failed to get container status \"93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7\": rpc error: code = NotFound desc = could not find container \"93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7\": container with ID starting with 93047a4f21f5a88cee28f592d1e9099632123bb54ba2581c3d39390fbcbe4dc7 not found: ID does not exist" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.352601 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.352665 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.352677 4771 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.361148 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f1f84431-dda1-43c4-8246-5742c470bf09" (UID: "f1f84431-dda1-43c4-8246-5742c470bf09"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.459571 4771 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1f84431-dda1-43c4-8246-5742c470bf09-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.467501 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.495624 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.521373 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 10:04:31 crc kubenswrapper[4771]: E1002 10:04:31.521964 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8072221f-4513-4f85-b4ea-f06d0b81214c" containerName="extract-utilities" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.521982 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8072221f-4513-4f85-b4ea-f06d0b81214c" containerName="extract-utilities" Oct 02 10:04:31 crc kubenswrapper[4771]: E1002 10:04:31.521997 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8072221f-4513-4f85-b4ea-f06d0b81214c" containerName="registry-server" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.522003 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8072221f-4513-4f85-b4ea-f06d0b81214c" containerName="registry-server" Oct 02 10:04:31 crc kubenswrapper[4771]: E1002 10:04:31.522015 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8072221f-4513-4f85-b4ea-f06d0b81214c" containerName="extract-content" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.522021 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8072221f-4513-4f85-b4ea-f06d0b81214c" containerName="extract-content" Oct 02 10:04:31 crc kubenswrapper[4771]: E1002 10:04:31.522050 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ccdd3d7-717c-4879-9155-41669900240b" containerName="init" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.522056 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ccdd3d7-717c-4879-9155-41669900240b" containerName="init" Oct 02 10:04:31 crc kubenswrapper[4771]: E1002 10:04:31.522066 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1f84431-dda1-43c4-8246-5742c470bf09" containerName="nova-api-log" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.522072 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1f84431-dda1-43c4-8246-5742c470bf09" containerName="nova-api-log" Oct 02 10:04:31 crc kubenswrapper[4771]: E1002 10:04:31.522085 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ccdd3d7-717c-4879-9155-41669900240b" containerName="dnsmasq-dns" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.522092 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ccdd3d7-717c-4879-9155-41669900240b" containerName="dnsmasq-dns" Oct 02 10:04:31 crc kubenswrapper[4771]: E1002 10:04:31.522107 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8da47e22-e6c2-4c9f-92f0-9ae49a82be69" containerName="nova-manage" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.522114 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8da47e22-e6c2-4c9f-92f0-9ae49a82be69" containerName="nova-manage" Oct 02 10:04:31 crc kubenswrapper[4771]: E1002 10:04:31.522135 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1f84431-dda1-43c4-8246-5742c470bf09" containerName="nova-api-api" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.522141 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1f84431-dda1-43c4-8246-5742c470bf09" containerName="nova-api-api" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.522373 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ccdd3d7-717c-4879-9155-41669900240b" containerName="dnsmasq-dns" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.522388 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1f84431-dda1-43c4-8246-5742c470bf09" containerName="nova-api-log" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.522423 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8072221f-4513-4f85-b4ea-f06d0b81214c" containerName="registry-server" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.522444 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8da47e22-e6c2-4c9f-92f0-9ae49a82be69" containerName="nova-manage" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.522459 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1f84431-dda1-43c4-8246-5742c470bf09" containerName="nova-api-api" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.523773 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.528693 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.528832 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.528872 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.541164 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.664507 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-public-tls-certs\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.664731 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-logs\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.664823 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-config-data\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.664883 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.666342 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.666423 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjgtr\" (UniqueName: \"kubernetes.io/projected/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-kube-api-access-pjgtr\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.702777 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1f84431-dda1-43c4-8246-5742c470bf09" path="/var/lib/kubelet/pods/f1f84431-dda1-43c4-8246-5742c470bf09/volumes" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.768871 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.770480 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.770804 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjgtr\" (UniqueName: \"kubernetes.io/projected/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-kube-api-access-pjgtr\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.771138 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-public-tls-certs\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.771252 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-logs\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.771396 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-config-data\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.773267 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.773752 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-logs\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.776864 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-config-data\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.777098 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-public-tls-certs\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.783725 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.801201 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjgtr\" (UniqueName: \"kubernetes.io/projected/40e3003d-6fb9-455d-b3d6-0bebf099b5d1-kube-api-access-pjgtr\") pod \"nova-api-0\" (UID: \"40e3003d-6fb9-455d-b3d6-0bebf099b5d1\") " pod="openstack/nova-api-0" Oct 02 10:04:31 crc kubenswrapper[4771]: I1002 10:04:31.844966 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:04:32 crc kubenswrapper[4771]: I1002 10:04:32.140035 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerName="nova-metadata-log" containerID="cri-o://0baa500f430ba2d9d5bebdb0339e4603a2dee13d24b5c04f29f73a1e780fad07" gracePeriod=30 Oct 02 10:04:32 crc kubenswrapper[4771]: I1002 10:04:32.140189 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerName="nova-metadata-metadata" containerID="cri-o://61d47adbfb336e1f32080454ef121b0ee51109ce466b0408832e451ee6ad66a8" gracePeriod=30 Oct 02 10:04:32 crc kubenswrapper[4771]: I1002 10:04:32.325093 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:04:32 crc kubenswrapper[4771]: E1002 10:04:32.393594 4771 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/866e18c5cd00b8b18692bc087068afc65504c5558337a094ba56a172f7eaf59d/diff" to get inode usage: stat /var/lib/containers/storage/overlay/866e18c5cd00b8b18692bc087068afc65504c5558337a094ba56a172f7eaf59d/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_dnsmasq-dns-9b86998b5-gphqv_7ccdd3d7-717c-4879-9155-41669900240b/dnsmasq-dns/0.log" to get inode usage: stat /var/log/pods/openstack_dnsmasq-dns-9b86998b5-gphqv_7ccdd3d7-717c-4879-9155-41669900240b/dnsmasq-dns/0.log: no such file or directory Oct 02 10:04:32 crc kubenswrapper[4771]: I1002 10:04:32.889943 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.064884 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-config-data\") pod \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\" (UID: \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\") " Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.065222 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nwx9\" (UniqueName: \"kubernetes.io/projected/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-kube-api-access-8nwx9\") pod \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\" (UID: \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\") " Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.065265 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-combined-ca-bundle\") pod \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\" (UID: \"d343eb2f-faff-47b8-aad6-9d24a3fc39c7\") " Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.071502 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-kube-api-access-8nwx9" (OuterVolumeSpecName: "kube-api-access-8nwx9") pod "d343eb2f-faff-47b8-aad6-9d24a3fc39c7" (UID: "d343eb2f-faff-47b8-aad6-9d24a3fc39c7"). InnerVolumeSpecName "kube-api-access-8nwx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.112618 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d343eb2f-faff-47b8-aad6-9d24a3fc39c7" (UID: "d343eb2f-faff-47b8-aad6-9d24a3fc39c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.119508 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-config-data" (OuterVolumeSpecName: "config-data") pod "d343eb2f-faff-47b8-aad6-9d24a3fc39c7" (UID: "d343eb2f-faff-47b8-aad6-9d24a3fc39c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.154149 4771 generic.go:334] "Generic (PLEG): container finished" podID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerID="0baa500f430ba2d9d5bebdb0339e4603a2dee13d24b5c04f29f73a1e780fad07" exitCode=143 Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.154234 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a17c7e4c-40e5-491f-bd4d-6e8a224c9421","Type":"ContainerDied","Data":"0baa500f430ba2d9d5bebdb0339e4603a2dee13d24b5c04f29f73a1e780fad07"} Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.156035 4771 generic.go:334] "Generic (PLEG): container finished" podID="d343eb2f-faff-47b8-aad6-9d24a3fc39c7" containerID="eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911" exitCode=0 Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.156073 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.156080 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d343eb2f-faff-47b8-aad6-9d24a3fc39c7","Type":"ContainerDied","Data":"eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911"} Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.156429 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d343eb2f-faff-47b8-aad6-9d24a3fc39c7","Type":"ContainerDied","Data":"b5e9ee1f4991d03e79d5ff8918074a64c44b2359ea7671b5765eb36f46c043bd"} Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.156468 4771 scope.go:117] "RemoveContainer" containerID="eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.158583 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"40e3003d-6fb9-455d-b3d6-0bebf099b5d1","Type":"ContainerStarted","Data":"0b68e927035cfe83541bd30edd5422c85913de465bcadaa3f504bffc30b51795"} Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.158715 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"40e3003d-6fb9-455d-b3d6-0bebf099b5d1","Type":"ContainerStarted","Data":"2e71e55b0f737fac88aad0b5af893849ce8fedbe747e8fac69fc7a19e1055b91"} Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.158841 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"40e3003d-6fb9-455d-b3d6-0bebf099b5d1","Type":"ContainerStarted","Data":"99bc50fe1f5f6951e2e3a69248259d9f9df3a8c80faf7f49d152b77f931426d4"} Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.168205 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nwx9\" (UniqueName: \"kubernetes.io/projected/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-kube-api-access-8nwx9\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.168505 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.168613 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d343eb2f-faff-47b8-aad6-9d24a3fc39c7-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.188582 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.188559999 podStartE2EDuration="2.188559999s" podCreationTimestamp="2025-10-02 10:04:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:04:33.184049982 +0000 UTC m=+1660.831735069" watchObservedRunningTime="2025-10-02 10:04:33.188559999 +0000 UTC m=+1660.836245066" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.194695 4771 scope.go:117] "RemoveContainer" containerID="eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911" Oct 02 10:04:33 crc kubenswrapper[4771]: E1002 10:04:33.195510 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911\": container with ID starting with eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911 not found: ID does not exist" containerID="eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.195570 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911"} err="failed to get container status \"eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911\": rpc error: code = NotFound desc = could not find container \"eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911\": container with ID starting with eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911 not found: ID does not exist" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.221553 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.235039 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.246631 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:04:33 crc kubenswrapper[4771]: E1002 10:04:33.247504 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d343eb2f-faff-47b8-aad6-9d24a3fc39c7" containerName="nova-scheduler-scheduler" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.247532 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d343eb2f-faff-47b8-aad6-9d24a3fc39c7" containerName="nova-scheduler-scheduler" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.247895 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="d343eb2f-faff-47b8-aad6-9d24a3fc39c7" containerName="nova-scheduler-scheduler" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.249380 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.257063 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.260007 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.376828 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j948\" (UniqueName: \"kubernetes.io/projected/8d385667-ed65-45c5-8d75-b299520569b5-kube-api-access-9j948\") pod \"nova-scheduler-0\" (UID: \"8d385667-ed65-45c5-8d75-b299520569b5\") " pod="openstack/nova-scheduler-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.377349 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d385667-ed65-45c5-8d75-b299520569b5-config-data\") pod \"nova-scheduler-0\" (UID: \"8d385667-ed65-45c5-8d75-b299520569b5\") " pod="openstack/nova-scheduler-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.377405 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d385667-ed65-45c5-8d75-b299520569b5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d385667-ed65-45c5-8d75-b299520569b5\") " pod="openstack/nova-scheduler-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.479708 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d385667-ed65-45c5-8d75-b299520569b5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d385667-ed65-45c5-8d75-b299520569b5\") " pod="openstack/nova-scheduler-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.479866 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j948\" (UniqueName: \"kubernetes.io/projected/8d385667-ed65-45c5-8d75-b299520569b5-kube-api-access-9j948\") pod \"nova-scheduler-0\" (UID: \"8d385667-ed65-45c5-8d75-b299520569b5\") " pod="openstack/nova-scheduler-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.479997 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d385667-ed65-45c5-8d75-b299520569b5-config-data\") pod \"nova-scheduler-0\" (UID: \"8d385667-ed65-45c5-8d75-b299520569b5\") " pod="openstack/nova-scheduler-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.484162 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d385667-ed65-45c5-8d75-b299520569b5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d385667-ed65-45c5-8d75-b299520569b5\") " pod="openstack/nova-scheduler-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.493866 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d385667-ed65-45c5-8d75-b299520569b5-config-data\") pod \"nova-scheduler-0\" (UID: \"8d385667-ed65-45c5-8d75-b299520569b5\") " pod="openstack/nova-scheduler-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.497289 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j948\" (UniqueName: \"kubernetes.io/projected/8d385667-ed65-45c5-8d75-b299520569b5-kube-api-access-9j948\") pod \"nova-scheduler-0\" (UID: \"8d385667-ed65-45c5-8d75-b299520569b5\") " pod="openstack/nova-scheduler-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.582554 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.737380 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d343eb2f-faff-47b8-aad6-9d24a3fc39c7" path="/var/lib/kubelet/pods/d343eb2f-faff-47b8-aad6-9d24a3fc39c7/volumes" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.928056 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:04:33 crc kubenswrapper[4771]: I1002 10:04:33.941341 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 02 10:04:34 crc kubenswrapper[4771]: I1002 10:04:34.090838 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:04:34 crc kubenswrapper[4771]: I1002 10:04:34.177735 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d385667-ed65-45c5-8d75-b299520569b5","Type":"ContainerStarted","Data":"7e1da86e857e83d2a8ea6b1145761cdcef1e4779ab41cba9208aa2a0ff1bd9f1"} Oct 02 10:04:35 crc kubenswrapper[4771]: I1002 10:04:35.194753 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d385667-ed65-45c5-8d75-b299520569b5","Type":"ContainerStarted","Data":"81ad83e20744533627a8344c34bcc3b8b487d6261e3067bfbf8808c0cc4c2794"} Oct 02 10:04:35 crc kubenswrapper[4771]: I1002 10:04:35.212165 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.212144602 podStartE2EDuration="2.212144602s" podCreationTimestamp="2025-10-02 10:04:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:04:35.210941311 +0000 UTC m=+1662.858626378" watchObservedRunningTime="2025-10-02 10:04:35.212144602 +0000 UTC m=+1662.859829659" Oct 02 10:04:35 crc kubenswrapper[4771]: I1002 10:04:35.293486 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.254:8775/\": read tcp 10.217.0.2:54696->10.217.0.254:8775: read: connection reset by peer" Oct 02 10:04:35 crc kubenswrapper[4771]: I1002 10:04:35.293577 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.254:8775/\": read tcp 10.217.0.2:54704->10.217.0.254:8775: read: connection reset by peer" Oct 02 10:04:35 crc kubenswrapper[4771]: W1002 10:04:35.499957 4771 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-conmon-01aeca9e1ffa450c46ed0ded04944aea8c92057279c8ac0632ac8e8fdcd1f883.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-conmon-01aeca9e1ffa450c46ed0ded04944aea8c92057279c8ac0632ac8e8fdcd1f883.scope: no such file or directory Oct 02 10:04:35 crc kubenswrapper[4771]: W1002 10:04:35.500033 4771 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-01aeca9e1ffa450c46ed0ded04944aea8c92057279c8ac0632ac8e8fdcd1f883.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-01aeca9e1ffa450c46ed0ded04944aea8c92057279c8ac0632ac8e8fdcd1f883.scope: no such file or directory Oct 02 10:04:35 crc kubenswrapper[4771]: W1002 10:04:35.500058 4771 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8072221f_4513_4f85_b4ea_f06d0b81214c.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8072221f_4513_4f85_b4ea_f06d0b81214c.slice: no such file or directory Oct 02 10:04:35 crc kubenswrapper[4771]: W1002 10:04:35.500104 4771 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-conmon-7a4a5dcedcdd2808566d1dece6c391df5b560ba0717684ab2acce35ea635a5d1.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-conmon-7a4a5dcedcdd2808566d1dece6c391df5b560ba0717684ab2acce35ea635a5d1.scope: no such file or directory Oct 02 10:04:35 crc kubenswrapper[4771]: W1002 10:04:35.500185 4771 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-7a4a5dcedcdd2808566d1dece6c391df5b560ba0717684ab2acce35ea635a5d1.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-7a4a5dcedcdd2808566d1dece6c391df5b560ba0717684ab2acce35ea635a5d1.scope: no such file or directory Oct 02 10:04:35 crc kubenswrapper[4771]: E1002 10:04:35.509769 4771 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/ffff44fd5e657580ef61f543ccda16b3f7042c10c08d8675525fe3c129eeed0a/diff" to get inode usage: stat /var/lib/containers/storage/overlay/ffff44fd5e657580ef61f543ccda16b3f7042c10c08d8675525fe3c129eeed0a/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_nova-cell1-novncproxy-0_ab155770-4f5d-43cf-a234-7c9394177ef6/nova-cell1-novncproxy-novncproxy/0.log" to get inode usage: stat /var/log/pods/openstack_nova-cell1-novncproxy-0_ab155770-4f5d-43cf-a234-7c9394177ef6/nova-cell1-novncproxy-novncproxy/0.log: no such file or directory Oct 02 10:04:35 crc kubenswrapper[4771]: W1002 10:04:35.518561 4771 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-conmon-7f020577297a4d53ef7fdb1e0a360574e80bc0959ace2d891f409dd0403c7541.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-conmon-7f020577297a4d53ef7fdb1e0a360574e80bc0959ace2d891f409dd0403c7541.scope: no such file or directory Oct 02 10:04:35 crc kubenswrapper[4771]: W1002 10:04:35.518608 4771 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-7f020577297a4d53ef7fdb1e0a360574e80bc0959ace2d891f409dd0403c7541.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-7f020577297a4d53ef7fdb1e0a360574e80bc0959ace2d891f409dd0403c7541.scope: no such file or directory Oct 02 10:04:35 crc kubenswrapper[4771]: W1002 10:04:35.522738 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd343eb2f_faff_47b8_aad6_9d24a3fc39c7.slice/crio-eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911.scope WatchSource:0}: Error finding container eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911: Status 404 returned error can't find the container with id eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911 Oct 02 10:04:35 crc kubenswrapper[4771]: E1002 10:04:35.650636 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff529351_3126_40ed_9d60_87959d520219.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-9467df9eaffc2ba3c84b6bce0d755e29d56809a3fee62731a1e3e535a9af9035.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-7f5f125117d1b1388a8503a71ec4c602610220b192068f6624d985dd54031912\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-d708a48848c28ecb797ca149a414b0ed2e22f15e3d4c604bdf2da2ca2b780207.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-da6a2166316f9f1ead80c45c13320c5f8cdf387ddf0ee787ebba09e521d4cf4e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-97fdbd831c00862cda41bbf77db00d97cea8c8d22204850b73ed21a0c5975289.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff529351_3126_40ed_9d60_87959d520219.slice/crio-abcc81ff8aac70a133b6b5adb36daf0b68556f0e59a78f41ba913285f95ceee0\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-conmon-da6a2166316f9f1ead80c45c13320c5f8cdf387ddf0ee787ebba09e521d4cf4e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-conmon-9467df9eaffc2ba3c84b6bce0d755e29d56809a3fee62731a1e3e535a9af9035.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-conmon-97fdbd831c00862cda41bbf77db00d97cea8c8d22204850b73ed21a0c5975289.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-conmon-d708a48848c28ecb797ca149a414b0ed2e22f15e3d4c604bdf2da2ca2b780207.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice\": RecentStats: unable to find data in memory cache]" Oct 02 10:04:35 crc kubenswrapper[4771]: E1002 10:04:35.650698 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-9467df9eaffc2ba3c84b6bce0d755e29d56809a3fee62731a1e3e535a9af9035.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd343eb2f_faff_47b8_aad6_9d24a3fc39c7.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-97fdbd831c00862cda41bbf77db00d97cea8c8d22204850b73ed21a0c5975289.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-78e04f9546fe73ae805651aa40ff364d3723d2decf17de1b2cba0a143d57cd81.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-da6a2166316f9f1ead80c45c13320c5f8cdf387ddf0ee787ebba09e521d4cf4e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-conmon-d708a48848c28ecb797ca149a414b0ed2e22f15e3d4c604bdf2da2ca2b780207.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ad3c2_55fb_4cd7_a16b_fbf614f3bca1.slice/crio-conmon-78e04f9546fe73ae805651aa40ff364d3723d2decf17de1b2cba0a143d57cd81.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd343eb2f_faff_47b8_aad6_9d24a3fc39c7.slice/crio-conmon-eebb92d87a84185a7e307840d4597896320beb221cca935c3f295bed50b16911.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-conmon-97fdbd831c00862cda41bbf77db00d97cea8c8d22204850b73ed21a0c5975289.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd343eb2f_faff_47b8_aad6_9d24a3fc39c7.slice/crio-b5e9ee1f4991d03e79d5ff8918074a64c44b2359ea7671b5765eb36f46c043bd\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-d708a48848c28ecb797ca149a414b0ed2e22f15e3d4c604bdf2da2ca2b780207.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-7f5f125117d1b1388a8503a71ec4c602610220b192068f6624d985dd54031912\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-conmon-9467df9eaffc2ba3c84b6bce0d755e29d56809a3fee62731a1e3e535a9af9035.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:04:35 crc kubenswrapper[4771]: E1002 10:04:35.650876 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-7f5f125117d1b1388a8503a71ec4c602610220b192068f6624d985dd54031912\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff529351_3126_40ed_9d60_87959d520219.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-d708a48848c28ecb797ca149a414b0ed2e22f15e3d4c604bdf2da2ca2b780207.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-conmon-97fdbd831c00862cda41bbf77db00d97cea8c8d22204850b73ed21a0c5975289.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-conmon-d708a48848c28ecb797ca149a414b0ed2e22f15e3d4c604bdf2da2ca2b780207.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-conmon-9467df9eaffc2ba3c84b6bce0d755e29d56809a3fee62731a1e3e535a9af9035.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-97fdbd831c00862cda41bbf77db00d97cea8c8d22204850b73ed21a0c5975289.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-9467df9eaffc2ba3c84b6bce0d755e29d56809a3fee62731a1e3e535a9af9035.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-da6a2166316f9f1ead80c45c13320c5f8cdf387ddf0ee787ebba09e521d4cf4e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice/crio-conmon-da6a2166316f9f1ead80c45c13320c5f8cdf387ddf0ee787ebba09e521d4cf4e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9158fae1_e129_47d9_af54_2105c7d79e41.slice\": RecentStats: unable to find data in memory cache]" Oct 02 10:04:35 crc kubenswrapper[4771]: E1002 10:04:35.923964 4771 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/77ab68abd35e4fc540c8957c061d4c0608de006be18cacff5d1ddc7fe7d47c09/diff" to get inode usage: stat /var/lib/containers/storage/overlay/77ab68abd35e4fc540c8957c061d4c0608de006be18cacff5d1ddc7fe7d47c09/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_nova-metadata-0_8660c7e8-08fa-49fc-ab03-f0346caedce4/nova-metadata-metadata/0.log" to get inode usage: stat /var/log/pods/openstack_nova-metadata-0_8660c7e8-08fa-49fc-ab03-f0346caedce4/nova-metadata-metadata/0.log: no such file or directory Oct 02 10:04:35 crc kubenswrapper[4771]: I1002 10:04:35.977224 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.144263 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-combined-ca-bundle\") pod \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.144436 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hn7pk\" (UniqueName: \"kubernetes.io/projected/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-kube-api-access-hn7pk\") pod \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.144489 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-nova-metadata-tls-certs\") pod \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.144571 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-logs\") pod \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.144605 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-config-data\") pod \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\" (UID: \"a17c7e4c-40e5-491f-bd4d-6e8a224c9421\") " Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.145588 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-logs" (OuterVolumeSpecName: "logs") pod "a17c7e4c-40e5-491f-bd4d-6e8a224c9421" (UID: "a17c7e4c-40e5-491f-bd4d-6e8a224c9421"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.151771 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-kube-api-access-hn7pk" (OuterVolumeSpecName: "kube-api-access-hn7pk") pod "a17c7e4c-40e5-491f-bd4d-6e8a224c9421" (UID: "a17c7e4c-40e5-491f-bd4d-6e8a224c9421"). InnerVolumeSpecName "kube-api-access-hn7pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.217942 4771 generic.go:334] "Generic (PLEG): container finished" podID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerID="7f020577297a4d53ef7fdb1e0a360574e80bc0959ace2d891f409dd0403c7541" exitCode=137 Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.217975 4771 generic.go:334] "Generic (PLEG): container finished" podID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerID="7a4a5dcedcdd2808566d1dece6c391df5b560ba0717684ab2acce35ea635a5d1" exitCode=137 Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.218017 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1","Type":"ContainerDied","Data":"7f020577297a4d53ef7fdb1e0a360574e80bc0959ace2d891f409dd0403c7541"} Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.218045 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1","Type":"ContainerDied","Data":"7a4a5dcedcdd2808566d1dece6c391df5b560ba0717684ab2acce35ea635a5d1"} Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.221280 4771 generic.go:334] "Generic (PLEG): container finished" podID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerID="61d47adbfb336e1f32080454ef121b0ee51109ce466b0408832e451ee6ad66a8" exitCode=0 Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.222670 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.222844 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a17c7e4c-40e5-491f-bd4d-6e8a224c9421","Type":"ContainerDied","Data":"61d47adbfb336e1f32080454ef121b0ee51109ce466b0408832e451ee6ad66a8"} Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.222910 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a17c7e4c-40e5-491f-bd4d-6e8a224c9421","Type":"ContainerDied","Data":"37db0e096ef5b1d568ed327b4432ca8ad5c9ced517a34620f11b77bb15b62835"} Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.222937 4771 scope.go:117] "RemoveContainer" containerID="61d47adbfb336e1f32080454ef121b0ee51109ce466b0408832e451ee6ad66a8" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.241664 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-config-data" (OuterVolumeSpecName: "config-data") pod "a17c7e4c-40e5-491f-bd4d-6e8a224c9421" (UID: "a17c7e4c-40e5-491f-bd4d-6e8a224c9421"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.250981 4771 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.260332 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.253411 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a17c7e4c-40e5-491f-bd4d-6e8a224c9421" (UID: "a17c7e4c-40e5-491f-bd4d-6e8a224c9421"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.260356 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hn7pk\" (UniqueName: \"kubernetes.io/projected/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-kube-api-access-hn7pk\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.253514 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a17c7e4c-40e5-491f-bd4d-6e8a224c9421" (UID: "a17c7e4c-40e5-491f-bd4d-6e8a224c9421"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.269308 4771 scope.go:117] "RemoveContainer" containerID="0baa500f430ba2d9d5bebdb0339e4603a2dee13d24b5c04f29f73a1e780fad07" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.301787 4771 scope.go:117] "RemoveContainer" containerID="61d47adbfb336e1f32080454ef121b0ee51109ce466b0408832e451ee6ad66a8" Oct 02 10:04:36 crc kubenswrapper[4771]: E1002 10:04:36.304828 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61d47adbfb336e1f32080454ef121b0ee51109ce466b0408832e451ee6ad66a8\": container with ID starting with 61d47adbfb336e1f32080454ef121b0ee51109ce466b0408832e451ee6ad66a8 not found: ID does not exist" containerID="61d47adbfb336e1f32080454ef121b0ee51109ce466b0408832e451ee6ad66a8" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.304863 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61d47adbfb336e1f32080454ef121b0ee51109ce466b0408832e451ee6ad66a8"} err="failed to get container status \"61d47adbfb336e1f32080454ef121b0ee51109ce466b0408832e451ee6ad66a8\": rpc error: code = NotFound desc = could not find container \"61d47adbfb336e1f32080454ef121b0ee51109ce466b0408832e451ee6ad66a8\": container with ID starting with 61d47adbfb336e1f32080454ef121b0ee51109ce466b0408832e451ee6ad66a8 not found: ID does not exist" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.304911 4771 scope.go:117] "RemoveContainer" containerID="0baa500f430ba2d9d5bebdb0339e4603a2dee13d24b5c04f29f73a1e780fad07" Oct 02 10:04:36 crc kubenswrapper[4771]: E1002 10:04:36.306982 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0baa500f430ba2d9d5bebdb0339e4603a2dee13d24b5c04f29f73a1e780fad07\": container with ID starting with 0baa500f430ba2d9d5bebdb0339e4603a2dee13d24b5c04f29f73a1e780fad07 not found: ID does not exist" containerID="0baa500f430ba2d9d5bebdb0339e4603a2dee13d24b5c04f29f73a1e780fad07" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.307008 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0baa500f430ba2d9d5bebdb0339e4603a2dee13d24b5c04f29f73a1e780fad07"} err="failed to get container status \"0baa500f430ba2d9d5bebdb0339e4603a2dee13d24b5c04f29f73a1e780fad07\": rpc error: code = NotFound desc = could not find container \"0baa500f430ba2d9d5bebdb0339e4603a2dee13d24b5c04f29f73a1e780fad07\": container with ID starting with 0baa500f430ba2d9d5bebdb0339e4603a2dee13d24b5c04f29f73a1e780fad07 not found: ID does not exist" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.362879 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.362920 4771 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a17c7e4c-40e5-491f-bd4d-6e8a224c9421-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.650235 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.691943 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.702995 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.721567 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:04:36 crc kubenswrapper[4771]: E1002 10:04:36.722244 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-api" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.722265 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-api" Oct 02 10:04:36 crc kubenswrapper[4771]: E1002 10:04:36.722287 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-evaluator" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.722293 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-evaluator" Oct 02 10:04:36 crc kubenswrapper[4771]: E1002 10:04:36.722306 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-notifier" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.722312 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-notifier" Oct 02 10:04:36 crc kubenswrapper[4771]: E1002 10:04:36.722334 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerName="nova-metadata-log" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.722341 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerName="nova-metadata-log" Oct 02 10:04:36 crc kubenswrapper[4771]: E1002 10:04:36.722378 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerName="nova-metadata-metadata" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.722385 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerName="nova-metadata-metadata" Oct 02 10:04:36 crc kubenswrapper[4771]: E1002 10:04:36.722399 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-listener" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.722406 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-listener" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.722671 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-evaluator" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.722705 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerName="nova-metadata-metadata" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.722720 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-notifier" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.722739 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-api" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.722766 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" containerName="aodh-listener" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.722774 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" containerName="nova-metadata-log" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.724336 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.736034 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.739959 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.754815 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.777323 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-scripts\") pod \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.777401 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-config-data\") pod \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.777459 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-combined-ca-bundle\") pod \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.777624 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r5xh\" (UniqueName: \"kubernetes.io/projected/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-kube-api-access-6r5xh\") pod \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\" (UID: \"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1\") " Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.778236 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-config-data\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.778315 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkrdf\" (UniqueName: \"kubernetes.io/projected/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-kube-api-access-zkrdf\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.778365 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-logs\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.778398 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.778430 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.797486 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-scripts" (OuterVolumeSpecName: "scripts") pod "6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" (UID: "6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.805530 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-kube-api-access-6r5xh" (OuterVolumeSpecName: "kube-api-access-6r5xh") pod "6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" (UID: "6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1"). InnerVolumeSpecName "kube-api-access-6r5xh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.880538 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-config-data\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.880637 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkrdf\" (UniqueName: \"kubernetes.io/projected/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-kube-api-access-zkrdf\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.880686 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-logs\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.880719 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.880750 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.884664 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r5xh\" (UniqueName: \"kubernetes.io/projected/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-kube-api-access-6r5xh\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.889847 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-logs\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.892194 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.907807 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.909735 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkrdf\" (UniqueName: \"kubernetes.io/projected/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-kube-api-access-zkrdf\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.911851 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.914784 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c2141e-6fa2-45fc-9379-9a43a9cc9e06-config-data\") pod \"nova-metadata-0\" (UID: \"87c2141e-6fa2-45fc-9379-9a43a9cc9e06\") " pod="openstack/nova-metadata-0" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.985500 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-config-data" (OuterVolumeSpecName: "config-data") pod "6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" (UID: "6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:36 crc kubenswrapper[4771]: I1002 10:04:36.996066 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.011220 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" (UID: "6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.054706 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.098968 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.236830 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1","Type":"ContainerDied","Data":"11bfaed686721cc97fe9e21fe0d67e4d51e2a58f8c2758858cce3c71b1df0022"} Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.236891 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.237218 4771 scope.go:117] "RemoveContainer" containerID="7f020577297a4d53ef7fdb1e0a360574e80bc0959ace2d891f409dd0403c7541" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.276011 4771 scope.go:117] "RemoveContainer" containerID="7a4a5dcedcdd2808566d1dece6c391df5b560ba0717684ab2acce35ea635a5d1" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.294899 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.325258 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.333547 4771 scope.go:117] "RemoveContainer" containerID="01aeca9e1ffa450c46ed0ded04944aea8c92057279c8ac0632ac8e8fdcd1f883" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.338178 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.341634 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.344860 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.344936 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.345059 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.345125 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-mnlz2" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.345171 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.354881 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.378646 4771 scope.go:117] "RemoveContainer" containerID="78e04f9546fe73ae805651aa40ff364d3723d2decf17de1b2cba0a143d57cd81" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.405525 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-internal-tls-certs\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.405606 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.405675 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf8s6\" (UniqueName: \"kubernetes.io/projected/8f62371d-1395-42e3-b0e8-ec5f3439441c-kube-api-access-cf8s6\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.405784 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-config-data\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.406015 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-scripts\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.406190 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-public-tls-certs\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.508670 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-config-data\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.508805 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-scripts\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.508855 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-public-tls-certs\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.508934 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-internal-tls-certs\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.508979 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.509032 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf8s6\" (UniqueName: \"kubernetes.io/projected/8f62371d-1395-42e3-b0e8-ec5f3439441c-kube-api-access-cf8s6\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.514658 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-public-tls-certs\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.515318 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.517024 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-config-data\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.524861 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-internal-tls-certs\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.525700 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-scripts\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.531189 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf8s6\" (UniqueName: \"kubernetes.io/projected/8f62371d-1395-42e3-b0e8-ec5f3439441c-kube-api-access-cf8s6\") pod \"aodh-0\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.609165 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.675962 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.703254 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1" path="/var/lib/kubelet/pods/6a5ad3c2-55fb-4cd7-a16b-fbf614f3bca1/volumes" Oct 02 10:04:37 crc kubenswrapper[4771]: I1002 10:04:37.704283 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a17c7e4c-40e5-491f-bd4d-6e8a224c9421" path="/var/lib/kubelet/pods/a17c7e4c-40e5-491f-bd4d-6e8a224c9421/volumes" Oct 02 10:04:38 crc kubenswrapper[4771]: I1002 10:04:38.269751 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 02 10:04:38 crc kubenswrapper[4771]: I1002 10:04:38.283613 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"87c2141e-6fa2-45fc-9379-9a43a9cc9e06","Type":"ContainerStarted","Data":"ca268fcb33967a9d9aab4b86d7facbb2d55ac38b0ca360526f4209c522cd6138"} Oct 02 10:04:38 crc kubenswrapper[4771]: I1002 10:04:38.283913 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"87c2141e-6fa2-45fc-9379-9a43a9cc9e06","Type":"ContainerStarted","Data":"6eaeef40a4783e7630ee489f369c3ab39a5445d83172fbbd43b1693ecb9c4af3"} Oct 02 10:04:38 crc kubenswrapper[4771]: I1002 10:04:38.582851 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 10:04:38 crc kubenswrapper[4771]: I1002 10:04:38.681927 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:04:38 crc kubenswrapper[4771]: E1002 10:04:38.682292 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:04:39 crc kubenswrapper[4771]: I1002 10:04:39.315652 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8f62371d-1395-42e3-b0e8-ec5f3439441c","Type":"ContainerStarted","Data":"581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8"} Oct 02 10:04:39 crc kubenswrapper[4771]: I1002 10:04:39.316142 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8f62371d-1395-42e3-b0e8-ec5f3439441c","Type":"ContainerStarted","Data":"87777b2e0efe794e8660b8f467b2688298544db6a9652f3a61193c0bb4972ada"} Oct 02 10:04:39 crc kubenswrapper[4771]: I1002 10:04:39.318649 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"87c2141e-6fa2-45fc-9379-9a43a9cc9e06","Type":"ContainerStarted","Data":"aa0d6c0858fc4725f1e2d28884118ab14d0c5c15a441832b820a1b7508f4e1f0"} Oct 02 10:04:39 crc kubenswrapper[4771]: I1002 10:04:39.342361 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.3423397919999998 podStartE2EDuration="3.342339792s" podCreationTimestamp="2025-10-02 10:04:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:04:39.33644845 +0000 UTC m=+1666.984133527" watchObservedRunningTime="2025-10-02 10:04:39.342339792 +0000 UTC m=+1666.990024849" Oct 02 10:04:40 crc kubenswrapper[4771]: I1002 10:04:40.337562 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8f62371d-1395-42e3-b0e8-ec5f3439441c","Type":"ContainerStarted","Data":"5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d"} Oct 02 10:04:41 crc kubenswrapper[4771]: I1002 10:04:41.351739 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8f62371d-1395-42e3-b0e8-ec5f3439441c","Type":"ContainerStarted","Data":"85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38"} Oct 02 10:04:41 crc kubenswrapper[4771]: W1002 10:04:41.691910 4771 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8da47e22_e6c2_4c9f_92f0_9ae49a82be69.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8da47e22_e6c2_4c9f_92f0_9ae49a82be69.slice: no such file or directory Oct 02 10:04:41 crc kubenswrapper[4771]: W1002 10:04:41.700668 4771 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1f84431_dda1_43c4_8246_5742c470bf09.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1f84431_dda1_43c4_8246_5742c470bf09.slice: no such file or directory Oct 02 10:04:41 crc kubenswrapper[4771]: I1002 10:04:41.846583 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:04:41 crc kubenswrapper[4771]: I1002 10:04:41.847994 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.056350 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.056390 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.397846 4771 generic.go:334] "Generic (PLEG): container finished" podID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerID="fb1f20c4441437bc04c904478af7a5c8637101ec7be58695bf54b56e0da77905" exitCode=137 Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.399718 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc","Type":"ContainerDied","Data":"fb1f20c4441437bc04c904478af7a5c8637101ec7be58695bf54b56e0da77905"} Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.406095 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8f62371d-1395-42e3-b0e8-ec5f3439441c","Type":"ContainerStarted","Data":"3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337"} Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.433397 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.369083282 podStartE2EDuration="5.433380149s" podCreationTimestamp="2025-10-02 10:04:37 +0000 UTC" firstStartedPulling="2025-10-02 10:04:38.296563259 +0000 UTC m=+1665.944248316" lastFinishedPulling="2025-10-02 10:04:41.360860116 +0000 UTC m=+1669.008545183" observedRunningTime="2025-10-02 10:04:42.430351701 +0000 UTC m=+1670.078036768" watchObservedRunningTime="2025-10-02 10:04:42.433380149 +0000 UTC m=+1670.081065216" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.530916 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.638086 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-sg-core-conf-yaml\") pod \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.638265 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-run-httpd\") pod \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.638303 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9k95\" (UniqueName: \"kubernetes.io/projected/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-kube-api-access-n9k95\") pod \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.638360 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-ceilometer-tls-certs\") pod \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.638428 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-config-data\") pod \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.638465 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-log-httpd\") pod \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.638550 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-scripts\") pod \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.638569 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-combined-ca-bundle\") pod \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\" (UID: \"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc\") " Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.640230 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" (UID: "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.647225 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" (UID: "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.655503 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-scripts" (OuterVolumeSpecName: "scripts") pod "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" (UID: "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.663694 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-kube-api-access-n9k95" (OuterVolumeSpecName: "kube-api-access-n9k95") pod "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" (UID: "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc"). InnerVolumeSpecName "kube-api-access-n9k95". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.741949 4771 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.742373 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.742387 4771 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.742402 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9k95\" (UniqueName: \"kubernetes.io/projected/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-kube-api-access-n9k95\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.747552 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" (UID: "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.762566 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" (UID: "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.834299 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" (UID: "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.844623 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.844654 4771 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.844666 4771 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.856415 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="40e3003d-6fb9-455d-b3d6-0bebf099b5d1" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.3:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.856499 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="40e3003d-6fb9-455d-b3d6-0bebf099b5d1" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.3:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.893195 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-config-data" (OuterVolumeSpecName: "config-data") pod "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" (UID: "a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:04:42 crc kubenswrapper[4771]: I1002 10:04:42.947301 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.418978 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc","Type":"ContainerDied","Data":"03dbf26878bc3888bff3165d8550636c52c701528edddf8412ef9cd9662d4c65"} Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.419055 4771 scope.go:117] "RemoveContainer" containerID="fb1f20c4441437bc04c904478af7a5c8637101ec7be58695bf54b56e0da77905" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.419060 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.453113 4771 scope.go:117] "RemoveContainer" containerID="153c2127da2cae292d5bd4b9c6c7e26eab0676376ad8f0818fd8b1c31dac3cfa" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.461748 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.483378 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.506393 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:04:43 crc kubenswrapper[4771]: E1002 10:04:43.507009 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="proxy-httpd" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.507027 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="proxy-httpd" Oct 02 10:04:43 crc kubenswrapper[4771]: E1002 10:04:43.507045 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="ceilometer-central-agent" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.507052 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="ceilometer-central-agent" Oct 02 10:04:43 crc kubenswrapper[4771]: E1002 10:04:43.507093 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="ceilometer-notification-agent" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.507099 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="ceilometer-notification-agent" Oct 02 10:04:43 crc kubenswrapper[4771]: E1002 10:04:43.507109 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="sg-core" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.507114 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="sg-core" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.507346 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="ceilometer-central-agent" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.507371 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="ceilometer-notification-agent" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.507387 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="proxy-httpd" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.507395 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" containerName="sg-core" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.509934 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.512959 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.513368 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.513506 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.519736 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.558587 4771 scope.go:117] "RemoveContainer" containerID="7f6fbfc15f008d4c36872c95ba2428af7688dd4f5bf413f1417113f5a1c98e0f" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.582785 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.585623 4771 scope.go:117] "RemoveContainer" containerID="3c1f6add4476e5f41d38ed17aa09b9cff2d9afe4a31f03a8086de9fdbf88837f" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.628512 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.663792 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz2bg\" (UniqueName: \"kubernetes.io/projected/41dff599-cbf7-42bf-b185-92d44bbe4a01-kube-api-access-cz2bg\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.663917 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dff599-cbf7-42bf-b185-92d44bbe4a01-run-httpd\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.664005 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.664041 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.664071 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.664174 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-config-data\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.664206 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dff599-cbf7-42bf-b185-92d44bbe4a01-log-httpd\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.664241 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-scripts\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.705324 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc" path="/var/lib/kubelet/pods/a0b1f3ba-ffcd-49c4-8f40-8b4ead9d25bc/volumes" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.766989 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.767049 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.767078 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.767150 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-config-data\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.767174 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dff599-cbf7-42bf-b185-92d44bbe4a01-log-httpd\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.767199 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-scripts\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.767289 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz2bg\" (UniqueName: \"kubernetes.io/projected/41dff599-cbf7-42bf-b185-92d44bbe4a01-kube-api-access-cz2bg\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.767351 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dff599-cbf7-42bf-b185-92d44bbe4a01-run-httpd\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.768001 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dff599-cbf7-42bf-b185-92d44bbe4a01-run-httpd\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.768059 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dff599-cbf7-42bf-b185-92d44bbe4a01-log-httpd\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.775773 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.775836 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-scripts\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.786065 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.786081 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.787648 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-config-data\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.797995 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz2bg\" (UniqueName: \"kubernetes.io/projected/41dff599-cbf7-42bf-b185-92d44bbe4a01-kube-api-access-cz2bg\") pod \"ceilometer-0\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " pod="openstack/ceilometer-0" Oct 02 10:04:43 crc kubenswrapper[4771]: I1002 10:04:43.859391 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:04:44 crc kubenswrapper[4771]: I1002 10:04:44.414470 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:04:44 crc kubenswrapper[4771]: I1002 10:04:44.504282 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 10:04:45 crc kubenswrapper[4771]: I1002 10:04:45.477399 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dff599-cbf7-42bf-b185-92d44bbe4a01","Type":"ContainerStarted","Data":"a83520dc70d439314917715cdcf80808c5c3d3089fab074ba27a4d35e7f1053b"} Oct 02 10:04:45 crc kubenswrapper[4771]: I1002 10:04:45.477766 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dff599-cbf7-42bf-b185-92d44bbe4a01","Type":"ContainerStarted","Data":"816e1ebf19623db42e11fa02e84d9d8f874efb319ab65f304a57082de5fb25d4"} Oct 02 10:04:46 crc kubenswrapper[4771]: I1002 10:04:46.492642 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dff599-cbf7-42bf-b185-92d44bbe4a01","Type":"ContainerStarted","Data":"15ad62f3a2c0cba36a2993cfc2f36e2d4883666d4646884335e278215844e67e"} Oct 02 10:04:47 crc kubenswrapper[4771]: I1002 10:04:47.055743 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 10:04:47 crc kubenswrapper[4771]: I1002 10:04:47.055870 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 10:04:47 crc kubenswrapper[4771]: I1002 10:04:47.516654 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dff599-cbf7-42bf-b185-92d44bbe4a01","Type":"ContainerStarted","Data":"4192eb38c3d7c47dbd5f19a4a9f6cfad30fd8e5b8f2aafd5c1ac21959873cc5d"} Oct 02 10:04:48 crc kubenswrapper[4771]: I1002 10:04:48.069573 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="87c2141e-6fa2-45fc-9379-9a43a9cc9e06" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.5:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:04:48 crc kubenswrapper[4771]: I1002 10:04:48.069600 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="87c2141e-6fa2-45fc-9379-9a43a9cc9e06" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.5:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 10:04:48 crc kubenswrapper[4771]: I1002 10:04:48.537755 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dff599-cbf7-42bf-b185-92d44bbe4a01","Type":"ContainerStarted","Data":"c689d2ca5485f23366ca643f74b034bae6592c2aa2b55dddb163cf1b16c13e69"} Oct 02 10:04:48 crc kubenswrapper[4771]: I1002 10:04:48.539557 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:04:48 crc kubenswrapper[4771]: I1002 10:04:48.564160 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.992619495 podStartE2EDuration="5.564113978s" podCreationTimestamp="2025-10-02 10:04:43 +0000 UTC" firstStartedPulling="2025-10-02 10:04:44.432763739 +0000 UTC m=+1672.080448816" lastFinishedPulling="2025-10-02 10:04:48.004258232 +0000 UTC m=+1675.651943299" observedRunningTime="2025-10-02 10:04:48.560257609 +0000 UTC m=+1676.207942676" watchObservedRunningTime="2025-10-02 10:04:48.564113978 +0000 UTC m=+1676.211799045" Oct 02 10:04:51 crc kubenswrapper[4771]: I1002 10:04:51.683832 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:04:51 crc kubenswrapper[4771]: E1002 10:04:51.687476 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:04:51 crc kubenswrapper[4771]: I1002 10:04:51.852159 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 10:04:51 crc kubenswrapper[4771]: I1002 10:04:51.853092 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 10:04:51 crc kubenswrapper[4771]: I1002 10:04:51.853840 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 10:04:51 crc kubenswrapper[4771]: I1002 10:04:51.854362 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 10:04:51 crc kubenswrapper[4771]: I1002 10:04:51.859800 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 10:04:51 crc kubenswrapper[4771]: I1002 10:04:51.866965 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 10:04:57 crc kubenswrapper[4771]: I1002 10:04:57.061965 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 10:04:57 crc kubenswrapper[4771]: I1002 10:04:57.065881 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 10:04:57 crc kubenswrapper[4771]: I1002 10:04:57.066821 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 10:04:57 crc kubenswrapper[4771]: I1002 10:04:57.696024 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 10:04:59 crc kubenswrapper[4771]: I1002 10:04:59.211160 4771 scope.go:117] "RemoveContainer" containerID="db107d781bec174e1922a04a1d846e1cb676be4c68153f675441a3c53dc84bdb" Oct 02 10:04:59 crc kubenswrapper[4771]: I1002 10:04:59.240243 4771 scope.go:117] "RemoveContainer" containerID="867e182039885b01a702368079a7be3979b5981c01662ee61cd415ca8df3eec1" Oct 02 10:04:59 crc kubenswrapper[4771]: I1002 10:04:59.265757 4771 scope.go:117] "RemoveContainer" containerID="97f487f824331b0f6a26a035064204af8d304d5c05bdbf68aa793206ac3d929c" Oct 02 10:04:59 crc kubenswrapper[4771]: I1002 10:04:59.293960 4771 scope.go:117] "RemoveContainer" containerID="841ad4e2563abbb9c7417b30557ee4a224a2f5f20713a6e3af69c1cad31829f6" Oct 02 10:04:59 crc kubenswrapper[4771]: I1002 10:04:59.351723 4771 scope.go:117] "RemoveContainer" containerID="9eae86c3847c6698cad06d23de888102374ae3d5ceeb3625d76966b6461733f0" Oct 02 10:04:59 crc kubenswrapper[4771]: I1002 10:04:59.411542 4771 scope.go:117] "RemoveContainer" containerID="4ff2ad23fc2d0fe2e7d27804d86888db3fdce232cd0c67272337d99907ba4e7c" Oct 02 10:04:59 crc kubenswrapper[4771]: I1002 10:04:59.444226 4771 scope.go:117] "RemoveContainer" containerID="f401f8e208415441627053abbaf61bd8fa3facefcad79e96174e420ca6f40b3c" Oct 02 10:04:59 crc kubenswrapper[4771]: I1002 10:04:59.505552 4771 scope.go:117] "RemoveContainer" containerID="93ff4861ff36ead6cbbbee3368d13a88352a9bd65803366637c7b6d1a8f68246" Oct 02 10:04:59 crc kubenswrapper[4771]: I1002 10:04:59.562448 4771 scope.go:117] "RemoveContainer" containerID="efa3ef697839c4982a1936f5891ff9ca59a9e10a47012a890cba3c5167d8ec50" Oct 02 10:04:59 crc kubenswrapper[4771]: I1002 10:04:59.613780 4771 scope.go:117] "RemoveContainer" containerID="4f8147b79e28ceeee6d5ed5931c7135f3876fa2bf3de5b81d07dd2f380427504" Oct 02 10:05:03 crc kubenswrapper[4771]: I1002 10:05:03.697139 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:05:03 crc kubenswrapper[4771]: E1002 10:05:03.698413 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:05:13 crc kubenswrapper[4771]: I1002 10:05:13.884043 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 10:05:18 crc kubenswrapper[4771]: I1002 10:05:18.682062 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:05:18 crc kubenswrapper[4771]: E1002 10:05:18.684524 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.144940 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-dbght"] Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.156607 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-dbght"] Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.246485 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-h99xg"] Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.248538 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-h99xg" Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.268701 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-h99xg"] Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.330700 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-config-data\") pod \"heat-db-sync-h99xg\" (UID: \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\") " pod="openstack/heat-db-sync-h99xg" Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.330850 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-combined-ca-bundle\") pod \"heat-db-sync-h99xg\" (UID: \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\") " pod="openstack/heat-db-sync-h99xg" Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.330883 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwlxb\" (UniqueName: \"kubernetes.io/projected/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-kube-api-access-qwlxb\") pod \"heat-db-sync-h99xg\" (UID: \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\") " pod="openstack/heat-db-sync-h99xg" Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.433279 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-config-data\") pod \"heat-db-sync-h99xg\" (UID: \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\") " pod="openstack/heat-db-sync-h99xg" Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.433439 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-combined-ca-bundle\") pod \"heat-db-sync-h99xg\" (UID: \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\") " pod="openstack/heat-db-sync-h99xg" Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.433475 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwlxb\" (UniqueName: \"kubernetes.io/projected/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-kube-api-access-qwlxb\") pod \"heat-db-sync-h99xg\" (UID: \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\") " pod="openstack/heat-db-sync-h99xg" Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.440248 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-config-data\") pod \"heat-db-sync-h99xg\" (UID: \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\") " pod="openstack/heat-db-sync-h99xg" Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.440836 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-combined-ca-bundle\") pod \"heat-db-sync-h99xg\" (UID: \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\") " pod="openstack/heat-db-sync-h99xg" Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.450901 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwlxb\" (UniqueName: \"kubernetes.io/projected/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-kube-api-access-qwlxb\") pod \"heat-db-sync-h99xg\" (UID: \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\") " pod="openstack/heat-db-sync-h99xg" Oct 02 10:05:26 crc kubenswrapper[4771]: I1002 10:05:26.573918 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-h99xg" Oct 02 10:05:27 crc kubenswrapper[4771]: I1002 10:05:27.124663 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-h99xg"] Oct 02 10:05:27 crc kubenswrapper[4771]: I1002 10:05:27.702824 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="042a2fb4-504c-4448-922b-48db5507e964" path="/var/lib/kubelet/pods/042a2fb4-504c-4448-922b-48db5507e964/volumes" Oct 02 10:05:28 crc kubenswrapper[4771]: I1002 10:05:28.086044 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-h99xg" event={"ID":"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02","Type":"ContainerStarted","Data":"a9ecbee9db07b00f02d9a94847eaebe2b1a470b1f8488350a133ae07b92bf51b"} Oct 02 10:05:28 crc kubenswrapper[4771]: I1002 10:05:28.606070 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:05:28 crc kubenswrapper[4771]: I1002 10:05:28.606954 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="sg-core" containerID="cri-o://4192eb38c3d7c47dbd5f19a4a9f6cfad30fd8e5b8f2aafd5c1ac21959873cc5d" gracePeriod=30 Oct 02 10:05:28 crc kubenswrapper[4771]: I1002 10:05:28.607187 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="proxy-httpd" containerID="cri-o://c689d2ca5485f23366ca643f74b034bae6592c2aa2b55dddb163cf1b16c13e69" gracePeriod=30 Oct 02 10:05:28 crc kubenswrapper[4771]: I1002 10:05:28.607184 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="ceilometer-notification-agent" containerID="cri-o://15ad62f3a2c0cba36a2993cfc2f36e2d4883666d4646884335e278215844e67e" gracePeriod=30 Oct 02 10:05:28 crc kubenswrapper[4771]: I1002 10:05:28.607266 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="ceilometer-central-agent" containerID="cri-o://a83520dc70d439314917715cdcf80808c5c3d3089fab074ba27a4d35e7f1053b" gracePeriod=30 Oct 02 10:05:28 crc kubenswrapper[4771]: I1002 10:05:28.773373 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 10:05:29 crc kubenswrapper[4771]: I1002 10:05:29.108683 4771 generic.go:334] "Generic (PLEG): container finished" podID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerID="c689d2ca5485f23366ca643f74b034bae6592c2aa2b55dddb163cf1b16c13e69" exitCode=0 Oct 02 10:05:29 crc kubenswrapper[4771]: I1002 10:05:29.108867 4771 generic.go:334] "Generic (PLEG): container finished" podID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerID="4192eb38c3d7c47dbd5f19a4a9f6cfad30fd8e5b8f2aafd5c1ac21959873cc5d" exitCode=2 Oct 02 10:05:29 crc kubenswrapper[4771]: I1002 10:05:29.108881 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dff599-cbf7-42bf-b185-92d44bbe4a01","Type":"ContainerDied","Data":"c689d2ca5485f23366ca643f74b034bae6592c2aa2b55dddb163cf1b16c13e69"} Oct 02 10:05:29 crc kubenswrapper[4771]: I1002 10:05:29.108953 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dff599-cbf7-42bf-b185-92d44bbe4a01","Type":"ContainerDied","Data":"4192eb38c3d7c47dbd5f19a4a9f6cfad30fd8e5b8f2aafd5c1ac21959873cc5d"} Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.041216 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.144792 4771 generic.go:334] "Generic (PLEG): container finished" podID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerID="15ad62f3a2c0cba36a2993cfc2f36e2d4883666d4646884335e278215844e67e" exitCode=0 Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.144834 4771 generic.go:334] "Generic (PLEG): container finished" podID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerID="a83520dc70d439314917715cdcf80808c5c3d3089fab074ba27a4d35e7f1053b" exitCode=0 Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.144860 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dff599-cbf7-42bf-b185-92d44bbe4a01","Type":"ContainerDied","Data":"15ad62f3a2c0cba36a2993cfc2f36e2d4883666d4646884335e278215844e67e"} Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.144896 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dff599-cbf7-42bf-b185-92d44bbe4a01","Type":"ContainerDied","Data":"a83520dc70d439314917715cdcf80808c5c3d3089fab074ba27a4d35e7f1053b"} Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.144908 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dff599-cbf7-42bf-b185-92d44bbe4a01","Type":"ContainerDied","Data":"816e1ebf19623db42e11fa02e84d9d8f874efb319ab65f304a57082de5fb25d4"} Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.144920 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="816e1ebf19623db42e11fa02e84d9d8f874efb319ab65f304a57082de5fb25d4" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.148421 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.241348 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dff599-cbf7-42bf-b185-92d44bbe4a01-log-httpd\") pod \"41dff599-cbf7-42bf-b185-92d44bbe4a01\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.241646 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cz2bg\" (UniqueName: \"kubernetes.io/projected/41dff599-cbf7-42bf-b185-92d44bbe4a01-kube-api-access-cz2bg\") pod \"41dff599-cbf7-42bf-b185-92d44bbe4a01\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.241690 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-scripts\") pod \"41dff599-cbf7-42bf-b185-92d44bbe4a01\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.241782 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-sg-core-conf-yaml\") pod \"41dff599-cbf7-42bf-b185-92d44bbe4a01\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.241810 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-combined-ca-bundle\") pod \"41dff599-cbf7-42bf-b185-92d44bbe4a01\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.241881 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dff599-cbf7-42bf-b185-92d44bbe4a01-run-httpd\") pod \"41dff599-cbf7-42bf-b185-92d44bbe4a01\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.241910 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-config-data\") pod \"41dff599-cbf7-42bf-b185-92d44bbe4a01\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.241942 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-ceilometer-tls-certs\") pod \"41dff599-cbf7-42bf-b185-92d44bbe4a01\" (UID: \"41dff599-cbf7-42bf-b185-92d44bbe4a01\") " Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.244036 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41dff599-cbf7-42bf-b185-92d44bbe4a01-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "41dff599-cbf7-42bf-b185-92d44bbe4a01" (UID: "41dff599-cbf7-42bf-b185-92d44bbe4a01"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.244458 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41dff599-cbf7-42bf-b185-92d44bbe4a01-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "41dff599-cbf7-42bf-b185-92d44bbe4a01" (UID: "41dff599-cbf7-42bf-b185-92d44bbe4a01"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.247768 4771 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dff599-cbf7-42bf-b185-92d44bbe4a01-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.247801 4771 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dff599-cbf7-42bf-b185-92d44bbe4a01-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.251549 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-scripts" (OuterVolumeSpecName: "scripts") pod "41dff599-cbf7-42bf-b185-92d44bbe4a01" (UID: "41dff599-cbf7-42bf-b185-92d44bbe4a01"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.281052 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41dff599-cbf7-42bf-b185-92d44bbe4a01-kube-api-access-cz2bg" (OuterVolumeSpecName: "kube-api-access-cz2bg") pod "41dff599-cbf7-42bf-b185-92d44bbe4a01" (UID: "41dff599-cbf7-42bf-b185-92d44bbe4a01"). InnerVolumeSpecName "kube-api-access-cz2bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.349995 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cz2bg\" (UniqueName: \"kubernetes.io/projected/41dff599-cbf7-42bf-b185-92d44bbe4a01-kube-api-access-cz2bg\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.350040 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.417537 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "41dff599-cbf7-42bf-b185-92d44bbe4a01" (UID: "41dff599-cbf7-42bf-b185-92d44bbe4a01"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.455946 4771 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.505376 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "41dff599-cbf7-42bf-b185-92d44bbe4a01" (UID: "41dff599-cbf7-42bf-b185-92d44bbe4a01"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.518404 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-config-data" (OuterVolumeSpecName: "config-data") pod "41dff599-cbf7-42bf-b185-92d44bbe4a01" (UID: "41dff599-cbf7-42bf-b185-92d44bbe4a01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.531735 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41dff599-cbf7-42bf-b185-92d44bbe4a01" (UID: "41dff599-cbf7-42bf-b185-92d44bbe4a01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.558642 4771 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.559047 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:30 crc kubenswrapper[4771]: I1002 10:05:30.559141 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41dff599-cbf7-42bf-b185-92d44bbe4a01-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.166723 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.219682 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.235202 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.254160 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:05:31 crc kubenswrapper[4771]: E1002 10:05:31.264660 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="ceilometer-notification-agent" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.264713 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="ceilometer-notification-agent" Oct 02 10:05:31 crc kubenswrapper[4771]: E1002 10:05:31.264745 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="sg-core" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.264757 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="sg-core" Oct 02 10:05:31 crc kubenswrapper[4771]: E1002 10:05:31.264828 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="proxy-httpd" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.264837 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="proxy-httpd" Oct 02 10:05:31 crc kubenswrapper[4771]: E1002 10:05:31.264883 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="ceilometer-central-agent" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.264890 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="ceilometer-central-agent" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.265327 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="proxy-httpd" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.265367 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="ceilometer-notification-agent" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.265385 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="ceilometer-central-agent" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.265402 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" containerName="sg-core" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.268059 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.277949 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.278177 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.278784 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.282380 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.384935 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-scripts\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.384980 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.385004 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.385081 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.385108 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53db4baf-fd23-4d39-a4c8-0019beae7703-log-httpd\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.385194 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-config-data\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.385223 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53db4baf-fd23-4d39-a4c8-0019beae7703-run-httpd\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.385246 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55t2c\" (UniqueName: \"kubernetes.io/projected/53db4baf-fd23-4d39-a4c8-0019beae7703-kube-api-access-55t2c\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.487119 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-config-data\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.487192 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53db4baf-fd23-4d39-a4c8-0019beae7703-run-httpd\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.487222 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55t2c\" (UniqueName: \"kubernetes.io/projected/53db4baf-fd23-4d39-a4c8-0019beae7703-kube-api-access-55t2c\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.487315 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-scripts\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.487336 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.487357 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.487437 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.487463 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53db4baf-fd23-4d39-a4c8-0019beae7703-log-httpd\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.487900 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53db4baf-fd23-4d39-a4c8-0019beae7703-log-httpd\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.487905 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53db4baf-fd23-4d39-a4c8-0019beae7703-run-httpd\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.492143 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.492284 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.492938 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-config-data\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.496372 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.511209 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55t2c\" (UniqueName: \"kubernetes.io/projected/53db4baf-fd23-4d39-a4c8-0019beae7703-kube-api-access-55t2c\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.527860 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53db4baf-fd23-4d39-a4c8-0019beae7703-scripts\") pod \"ceilometer-0\" (UID: \"53db4baf-fd23-4d39-a4c8-0019beae7703\") " pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.597774 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.689035 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:05:31 crc kubenswrapper[4771]: E1002 10:05:31.689786 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:05:31 crc kubenswrapper[4771]: I1002 10:05:31.709671 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41dff599-cbf7-42bf-b185-92d44bbe4a01" path="/var/lib/kubelet/pods/41dff599-cbf7-42bf-b185-92d44bbe4a01/volumes" Oct 02 10:05:32 crc kubenswrapper[4771]: I1002 10:05:32.327786 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:05:33 crc kubenswrapper[4771]: I1002 10:05:33.200585 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53db4baf-fd23-4d39-a4c8-0019beae7703","Type":"ContainerStarted","Data":"74fe517dd50ff4bcabf08e4334e3df2563321aeec12eeb837b5c025a715acd26"} Oct 02 10:05:34 crc kubenswrapper[4771]: I1002 10:05:34.235903 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="ebc075f5-e2ba-4789-a38e-4065fe661ada" containerName="rabbitmq" containerID="cri-o://82fe7b3800926b8bc7dfa8813efaf7f0dd9c638720dde9673e1bf775cfa6e1dd" gracePeriod=604795 Oct 02 10:05:35 crc kubenswrapper[4771]: I1002 10:05:35.265624 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="b7dad196-4ebc-43e0-b850-0820f2b072c9" containerName="rabbitmq" containerID="cri-o://600abb9d49fe9d1c7f960409891f3dd4c17c8d27d72e596b2ef8946d71c7edff" gracePeriod=604795 Oct 02 10:05:42 crc kubenswrapper[4771]: I1002 10:05:42.361536 4771 generic.go:334] "Generic (PLEG): container finished" podID="ebc075f5-e2ba-4789-a38e-4065fe661ada" containerID="82fe7b3800926b8bc7dfa8813efaf7f0dd9c638720dde9673e1bf775cfa6e1dd" exitCode=0 Oct 02 10:05:42 crc kubenswrapper[4771]: I1002 10:05:42.361636 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ebc075f5-e2ba-4789-a38e-4065fe661ada","Type":"ContainerDied","Data":"82fe7b3800926b8bc7dfa8813efaf7f0dd9c638720dde9673e1bf775cfa6e1dd"} Oct 02 10:05:42 crc kubenswrapper[4771]: I1002 10:05:42.364756 4771 generic.go:334] "Generic (PLEG): container finished" podID="b7dad196-4ebc-43e0-b850-0820f2b072c9" containerID="600abb9d49fe9d1c7f960409891f3dd4c17c8d27d72e596b2ef8946d71c7edff" exitCode=0 Oct 02 10:05:42 crc kubenswrapper[4771]: I1002 10:05:42.364783 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7dad196-4ebc-43e0-b850-0820f2b072c9","Type":"ContainerDied","Data":"600abb9d49fe9d1c7f960409891f3dd4c17c8d27d72e596b2ef8946d71c7edff"} Oct 02 10:05:44 crc kubenswrapper[4771]: I1002 10:05:44.682498 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:05:44 crc kubenswrapper[4771]: E1002 10:05:44.684545 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.374553 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-dcwpt"] Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.422755 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-dcwpt"] Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.422930 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.427437 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="b7dad196-4ebc-43e0-b850-0820f2b072c9" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.131:5671: i/o timeout" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.432253 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.492080 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-config\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.496301 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.496761 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.498056 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.498953 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.499141 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hp6d\" (UniqueName: \"kubernetes.io/projected/4494bb18-da2b-4304-84a9-d6e8709797b8-kube-api-access-9hp6d\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.499284 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.501060 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-dcwpt"] Oct 02 10:05:48 crc kubenswrapper[4771]: E1002 10:05:48.504097 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-9hp6d openstack-edpm-ipam ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" podUID="4494bb18-da2b-4304-84a9-d6e8709797b8" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.545664 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-vzhpr"] Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.548321 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.564839 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-vzhpr"] Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.602359 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-config\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.602429 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.602519 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.602579 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.602637 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.602674 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.602725 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.602775 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7759\" (UniqueName: \"kubernetes.io/projected/faa0b798-d3ee-4469-b60b-15c77ecf1753-kube-api-access-v7759\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.602817 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hp6d\" (UniqueName: \"kubernetes.io/projected/4494bb18-da2b-4304-84a9-d6e8709797b8-kube-api-access-9hp6d\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.602875 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.602938 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.602969 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.602993 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-config\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.603237 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.604298 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-config\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.604671 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.604682 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.605787 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.606475 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.606512 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.630844 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hp6d\" (UniqueName: \"kubernetes.io/projected/4494bb18-da2b-4304-84a9-d6e8709797b8-kube-api-access-9hp6d\") pod \"dnsmasq-dns-7d84b4d45c-dcwpt\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.705791 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.706176 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-config\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.706262 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.706358 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.706378 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7759\" (UniqueName: \"kubernetes.io/projected/faa0b798-d3ee-4469-b60b-15c77ecf1753-kube-api-access-v7759\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.706474 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.706498 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.707541 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.708300 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-config\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.708633 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.709007 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.709366 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.709557 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/faa0b798-d3ee-4469-b60b-15c77ecf1753-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.735712 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7759\" (UniqueName: \"kubernetes.io/projected/faa0b798-d3ee-4469-b60b-15c77ecf1753-kube-api-access-v7759\") pod \"dnsmasq-dns-6f6df4f56c-vzhpr\" (UID: \"faa0b798-d3ee-4469-b60b-15c77ecf1753\") " pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.890477 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:48 crc kubenswrapper[4771]: I1002 10:05:48.925048 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="ebc075f5-e2ba-4789-a38e-4065fe661ada" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.132:5671: i/o timeout" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.500272 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.524592 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.633348 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-ovsdbserver-sb\") pod \"4494bb18-da2b-4304-84a9-d6e8709797b8\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.633527 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-dns-swift-storage-0\") pod \"4494bb18-da2b-4304-84a9-d6e8709797b8\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.633567 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-openstack-edpm-ipam\") pod \"4494bb18-da2b-4304-84a9-d6e8709797b8\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.633595 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hp6d\" (UniqueName: \"kubernetes.io/projected/4494bb18-da2b-4304-84a9-d6e8709797b8-kube-api-access-9hp6d\") pod \"4494bb18-da2b-4304-84a9-d6e8709797b8\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.633796 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-config\") pod \"4494bb18-da2b-4304-84a9-d6e8709797b8\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.633857 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-ovsdbserver-nb\") pod \"4494bb18-da2b-4304-84a9-d6e8709797b8\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.633887 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-dns-svc\") pod \"4494bb18-da2b-4304-84a9-d6e8709797b8\" (UID: \"4494bb18-da2b-4304-84a9-d6e8709797b8\") " Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.634063 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4494bb18-da2b-4304-84a9-d6e8709797b8" (UID: "4494bb18-da2b-4304-84a9-d6e8709797b8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.634412 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4494bb18-da2b-4304-84a9-d6e8709797b8" (UID: "4494bb18-da2b-4304-84a9-d6e8709797b8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.634540 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-config" (OuterVolumeSpecName: "config") pod "4494bb18-da2b-4304-84a9-d6e8709797b8" (UID: "4494bb18-da2b-4304-84a9-d6e8709797b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.634555 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4494bb18-da2b-4304-84a9-d6e8709797b8" (UID: "4494bb18-da2b-4304-84a9-d6e8709797b8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.635091 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "4494bb18-da2b-4304-84a9-d6e8709797b8" (UID: "4494bb18-da2b-4304-84a9-d6e8709797b8"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.635989 4771 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.636013 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.636026 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.636038 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.636048 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.637010 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4494bb18-da2b-4304-84a9-d6e8709797b8" (UID: "4494bb18-da2b-4304-84a9-d6e8709797b8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.639680 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4494bb18-da2b-4304-84a9-d6e8709797b8-kube-api-access-9hp6d" (OuterVolumeSpecName: "kube-api-access-9hp6d") pod "4494bb18-da2b-4304-84a9-d6e8709797b8" (UID: "4494bb18-da2b-4304-84a9-d6e8709797b8"). InnerVolumeSpecName "kube-api-access-9hp6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.739033 4771 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4494bb18-da2b-4304-84a9-d6e8709797b8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:49 crc kubenswrapper[4771]: I1002 10:05:49.739378 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hp6d\" (UniqueName: \"kubernetes.io/projected/4494bb18-da2b-4304-84a9-d6e8709797b8-kube-api-access-9hp6d\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:50 crc kubenswrapper[4771]: I1002 10:05:50.517072 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-dcwpt" Oct 02 10:05:50 crc kubenswrapper[4771]: I1002 10:05:50.586381 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-dcwpt"] Oct 02 10:05:50 crc kubenswrapper[4771]: I1002 10:05:50.599109 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-dcwpt"] Oct 02 10:05:51 crc kubenswrapper[4771]: I1002 10:05:51.695176 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4494bb18-da2b-4304-84a9-d6e8709797b8" path="/var/lib/kubelet/pods/4494bb18-da2b-4304-84a9-d6e8709797b8/volumes" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.413608 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.428702 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444015 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hz5b\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-kube-api-access-4hz5b\") pod \"b7dad196-4ebc-43e0-b850-0820f2b072c9\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444113 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ebc075f5-e2ba-4789-a38e-4065fe661ada-erlang-cookie-secret\") pod \"ebc075f5-e2ba-4789-a38e-4065fe661ada\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444161 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-erlang-cookie\") pod \"ebc075f5-e2ba-4789-a38e-4065fe661ada\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444211 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7dad196-4ebc-43e0-b850-0820f2b072c9-erlang-cookie-secret\") pod \"b7dad196-4ebc-43e0-b850-0820f2b072c9\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444246 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"b7dad196-4ebc-43e0-b850-0820f2b072c9\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444295 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ebc075f5-e2ba-4789-a38e-4065fe661ada\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444379 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-confd\") pod \"ebc075f5-e2ba-4789-a38e-4065fe661ada\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444478 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-confd\") pod \"b7dad196-4ebc-43e0-b850-0820f2b072c9\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444500 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-tls\") pod \"ebc075f5-e2ba-4789-a38e-4065fe661ada\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444550 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-plugins\") pod \"b7dad196-4ebc-43e0-b850-0820f2b072c9\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444585 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-tls\") pod \"b7dad196-4ebc-43e0-b850-0820f2b072c9\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444620 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-config-data\") pod \"ebc075f5-e2ba-4789-a38e-4065fe661ada\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444648 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ebc075f5-e2ba-4789-a38e-4065fe661ada-pod-info\") pod \"ebc075f5-e2ba-4789-a38e-4065fe661ada\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444696 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-server-conf\") pod \"ebc075f5-e2ba-4789-a38e-4065fe661ada\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444768 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnslr\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-kube-api-access-qnslr\") pod \"ebc075f5-e2ba-4789-a38e-4065fe661ada\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444816 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-erlang-cookie\") pod \"b7dad196-4ebc-43e0-b850-0820f2b072c9\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444879 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-config-data\") pod \"b7dad196-4ebc-43e0-b850-0820f2b072c9\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.444915 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7dad196-4ebc-43e0-b850-0820f2b072c9-pod-info\") pod \"b7dad196-4ebc-43e0-b850-0820f2b072c9\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.445093 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-plugins\") pod \"ebc075f5-e2ba-4789-a38e-4065fe661ada\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.445154 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-plugins-conf\") pod \"b7dad196-4ebc-43e0-b850-0820f2b072c9\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.445176 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-plugins-conf\") pod \"ebc075f5-e2ba-4789-a38e-4065fe661ada\" (UID: \"ebc075f5-e2ba-4789-a38e-4065fe661ada\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.445262 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-server-conf\") pod \"b7dad196-4ebc-43e0-b850-0820f2b072c9\" (UID: \"b7dad196-4ebc-43e0-b850-0820f2b072c9\") " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.445331 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ebc075f5-e2ba-4789-a38e-4065fe661ada" (UID: "ebc075f5-e2ba-4789-a38e-4065fe661ada"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.446310 4771 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.451073 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b7dad196-4ebc-43e0-b850-0820f2b072c9" (UID: "b7dad196-4ebc-43e0-b850-0820f2b072c9"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.455740 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b7dad196-4ebc-43e0-b850-0820f2b072c9" (UID: "b7dad196-4ebc-43e0-b850-0820f2b072c9"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.456485 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ebc075f5-e2ba-4789-a38e-4065fe661ada" (UID: "ebc075f5-e2ba-4789-a38e-4065fe661ada"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.461049 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b7dad196-4ebc-43e0-b850-0820f2b072c9" (UID: "b7dad196-4ebc-43e0-b850-0820f2b072c9"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.461571 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b7dad196-4ebc-43e0-b850-0820f2b072c9-pod-info" (OuterVolumeSpecName: "pod-info") pod "b7dad196-4ebc-43e0-b850-0820f2b072c9" (UID: "b7dad196-4ebc-43e0-b850-0820f2b072c9"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.462068 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ebc075f5-e2ba-4789-a38e-4065fe661ada" (UID: "ebc075f5-e2ba-4789-a38e-4065fe661ada"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.470216 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "ebc075f5-e2ba-4789-a38e-4065fe661ada" (UID: "ebc075f5-e2ba-4789-a38e-4065fe661ada"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.470838 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b7dad196-4ebc-43e0-b850-0820f2b072c9" (UID: "b7dad196-4ebc-43e0-b850-0820f2b072c9"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.482983 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-kube-api-access-qnslr" (OuterVolumeSpecName: "kube-api-access-qnslr") pod "ebc075f5-e2ba-4789-a38e-4065fe661ada" (UID: "ebc075f5-e2ba-4789-a38e-4065fe661ada"). InnerVolumeSpecName "kube-api-access-qnslr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.505441 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-config-data" (OuterVolumeSpecName: "config-data") pod "b7dad196-4ebc-43e0-b850-0820f2b072c9" (UID: "b7dad196-4ebc-43e0-b850-0820f2b072c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.544077 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-kube-api-access-4hz5b" (OuterVolumeSpecName: "kube-api-access-4hz5b") pod "b7dad196-4ebc-43e0-b850-0820f2b072c9" (UID: "b7dad196-4ebc-43e0-b850-0820f2b072c9"). InnerVolumeSpecName "kube-api-access-4hz5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.544295 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "b7dad196-4ebc-43e0-b850-0820f2b072c9" (UID: "b7dad196-4ebc-43e0-b850-0820f2b072c9"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.544877 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7dad196-4ebc-43e0-b850-0820f2b072c9-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b7dad196-4ebc-43e0-b850-0820f2b072c9" (UID: "b7dad196-4ebc-43e0-b850-0820f2b072c9"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.545349 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ebc075f5-e2ba-4789-a38e-4065fe661ada-pod-info" (OuterVolumeSpecName: "pod-info") pod "ebc075f5-e2ba-4789-a38e-4065fe661ada" (UID: "ebc075f5-e2ba-4789-a38e-4065fe661ada"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.546027 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebc075f5-e2ba-4789-a38e-4065fe661ada-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ebc075f5-e2ba-4789-a38e-4065fe661ada" (UID: "ebc075f5-e2ba-4789-a38e-4065fe661ada"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552564 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hz5b\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-kube-api-access-4hz5b\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552608 4771 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ebc075f5-e2ba-4789-a38e-4065fe661ada-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552621 4771 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b7dad196-4ebc-43e0-b850-0820f2b072c9-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552659 4771 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552679 4771 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552692 4771 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552703 4771 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552714 4771 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ebc075f5-e2ba-4789-a38e-4065fe661ada-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552725 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnslr\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-kube-api-access-qnslr\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552736 4771 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552748 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552759 4771 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b7dad196-4ebc-43e0-b850-0820f2b072c9-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552770 4771 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.552784 4771 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.553023 4771 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.561045 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ebc075f5-e2ba-4789-a38e-4065fe661ada" (UID: "ebc075f5-e2ba-4789-a38e-4065fe661ada"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.565237 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-config-data" (OuterVolumeSpecName: "config-data") pod "ebc075f5-e2ba-4789-a38e-4065fe661ada" (UID: "ebc075f5-e2ba-4789-a38e-4065fe661ada"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.582446 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-server-conf" (OuterVolumeSpecName: "server-conf") pod "b7dad196-4ebc-43e0-b850-0820f2b072c9" (UID: "b7dad196-4ebc-43e0-b850-0820f2b072c9"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.582734 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b7dad196-4ebc-43e0-b850-0820f2b072c9","Type":"ContainerDied","Data":"7bd65e96cfb2476a195caec04d490921c0aac5fba96d01ba8a91363fb21bcb5a"} Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.582790 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.582810 4771 scope.go:117] "RemoveContainer" containerID="600abb9d49fe9d1c7f960409891f3dd4c17c8d27d72e596b2ef8946d71c7edff" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.591442 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ebc075f5-e2ba-4789-a38e-4065fe661ada","Type":"ContainerDied","Data":"198faa5123c5bbbb5493a576910da7893b55dcc83131eb1be938faa0e81b0ab2"} Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.591575 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.601579 4771 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.616055 4771 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.633980 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-server-conf" (OuterVolumeSpecName: "server-conf") pod "ebc075f5-e2ba-4789-a38e-4065fe661ada" (UID: "ebc075f5-e2ba-4789-a38e-4065fe661ada"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.662316 4771 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.662359 4771 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.662370 4771 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.662380 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.662390 4771 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ebc075f5-e2ba-4789-a38e-4065fe661ada-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.662398 4771 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b7dad196-4ebc-43e0-b850-0820f2b072c9-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.723524 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ebc075f5-e2ba-4789-a38e-4065fe661ada" (UID: "ebc075f5-e2ba-4789-a38e-4065fe661ada"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.765165 4771 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ebc075f5-e2ba-4789-a38e-4065fe661ada-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.798271 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b7dad196-4ebc-43e0-b850-0820f2b072c9" (UID: "b7dad196-4ebc-43e0-b850-0820f2b072c9"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.872522 4771 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b7dad196-4ebc-43e0-b850-0820f2b072c9-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.924227 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.936502 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.951300 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 10:05:53 crc kubenswrapper[4771]: I1002 10:05:53.969330 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.001254 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 10:05:54 crc kubenswrapper[4771]: E1002 10:05:54.001838 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7dad196-4ebc-43e0-b850-0820f2b072c9" containerName="setup-container" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.001867 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7dad196-4ebc-43e0-b850-0820f2b072c9" containerName="setup-container" Oct 02 10:05:54 crc kubenswrapper[4771]: E1002 10:05:54.001890 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebc075f5-e2ba-4789-a38e-4065fe661ada" containerName="setup-container" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.001896 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebc075f5-e2ba-4789-a38e-4065fe661ada" containerName="setup-container" Oct 02 10:05:54 crc kubenswrapper[4771]: E1002 10:05:54.001906 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebc075f5-e2ba-4789-a38e-4065fe661ada" containerName="rabbitmq" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.001914 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebc075f5-e2ba-4789-a38e-4065fe661ada" containerName="rabbitmq" Oct 02 10:05:54 crc kubenswrapper[4771]: E1002 10:05:54.002030 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7dad196-4ebc-43e0-b850-0820f2b072c9" containerName="rabbitmq" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.002038 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7dad196-4ebc-43e0-b850-0820f2b072c9" containerName="rabbitmq" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.002297 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7dad196-4ebc-43e0-b850-0820f2b072c9" containerName="rabbitmq" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.002328 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebc075f5-e2ba-4789-a38e-4065fe661ada" containerName="rabbitmq" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.014225 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.019751 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.020313 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.020357 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.020458 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.020545 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.020616 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-v5mng" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.020703 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.034550 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.073250 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.078124 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.082268 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-kk7ht" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.082544 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.082912 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.083108 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.083426 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.083631 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.083805 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.087449 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07281b0c-ff0a-4195-9ddb-521645890aee-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.087621 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07281b0c-ff0a-4195-9ddb-521645890aee-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.087668 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07281b0c-ff0a-4195-9ddb-521645890aee-server-conf\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.087703 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.087781 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07281b0c-ff0a-4195-9ddb-521645890aee-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.087806 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07281b0c-ff0a-4195-9ddb-521645890aee-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.087838 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07281b0c-ff0a-4195-9ddb-521645890aee-pod-info\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.087877 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07281b0c-ff0a-4195-9ddb-521645890aee-config-data\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.087905 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xwkm\" (UniqueName: \"kubernetes.io/projected/07281b0c-ff0a-4195-9ddb-521645890aee-kube-api-access-6xwkm\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.087939 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07281b0c-ff0a-4195-9ddb-521645890aee-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.087974 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07281b0c-ff0a-4195-9ddb-521645890aee-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.110753 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.189657 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07281b0c-ff0a-4195-9ddb-521645890aee-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.189925 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07281b0c-ff0a-4195-9ddb-521645890aee-server-conf\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.190037 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.190427 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07281b0c-ff0a-4195-9ddb-521645890aee-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.190550 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07281b0c-ff0a-4195-9ddb-521645890aee-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.190211 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07281b0c-ff0a-4195-9ddb-521645890aee-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.190692 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07281b0c-ff0a-4195-9ddb-521645890aee-pod-info\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.190854 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07281b0c-ff0a-4195-9ddb-521645890aee-config-data\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.190907 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xwkm\" (UniqueName: \"kubernetes.io/projected/07281b0c-ff0a-4195-9ddb-521645890aee-kube-api-access-6xwkm\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.190958 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07281b0c-ff0a-4195-9ddb-521645890aee-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.191021 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07281b0c-ff0a-4195-9ddb-521645890aee-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.191107 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07281b0c-ff0a-4195-9ddb-521645890aee-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.191437 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07281b0c-ff0a-4195-9ddb-521645890aee-server-conf\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.191546 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07281b0c-ff0a-4195-9ddb-521645890aee-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.191663 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.191988 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07281b0c-ff0a-4195-9ddb-521645890aee-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.192581 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07281b0c-ff0a-4195-9ddb-521645890aee-config-data\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.196245 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07281b0c-ff0a-4195-9ddb-521645890aee-pod-info\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.196546 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07281b0c-ff0a-4195-9ddb-521645890aee-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.200303 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07281b0c-ff0a-4195-9ddb-521645890aee-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.200523 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07281b0c-ff0a-4195-9ddb-521645890aee-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.209364 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xwkm\" (UniqueName: \"kubernetes.io/projected/07281b0c-ff0a-4195-9ddb-521645890aee-kube-api-access-6xwkm\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.242664 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"07281b0c-ff0a-4195-9ddb-521645890aee\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.293893 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zvxq\" (UniqueName: \"kubernetes.io/projected/3fa19dbc-a10c-4863-bf25-035f9d6579e9-kube-api-access-9zvxq\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.293950 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3fa19dbc-a10c-4863-bf25-035f9d6579e9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.293989 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3fa19dbc-a10c-4863-bf25-035f9d6579e9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.294018 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3fa19dbc-a10c-4863-bf25-035f9d6579e9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.294046 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3fa19dbc-a10c-4863-bf25-035f9d6579e9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.294060 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3fa19dbc-a10c-4863-bf25-035f9d6579e9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.294077 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3fa19dbc-a10c-4863-bf25-035f9d6579e9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.294149 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3fa19dbc-a10c-4863-bf25-035f9d6579e9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.294239 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3fa19dbc-a10c-4863-bf25-035f9d6579e9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.294577 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.294628 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3fa19dbc-a10c-4863-bf25-035f9d6579e9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: E1002 10:05:54.339211 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Oct 02 10:05:54 crc kubenswrapper[4771]: E1002 10:05:54.339287 4771 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Oct 02 10:05:54 crc kubenswrapper[4771]: E1002 10:05:54.339471 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n646h698hc8h5cbh56ch5cbh664h685hbch589h89h5dhb7h5b5h56bh568h67bh96h565h59h98h654h579hbch55bh7fh5d8h7h5cch5b7h66chbcq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-55t2c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(53db4baf-fd23-4d39-a4c8-0019beae7703): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.348589 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.397377 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zvxq\" (UniqueName: \"kubernetes.io/projected/3fa19dbc-a10c-4863-bf25-035f9d6579e9-kube-api-access-9zvxq\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.397446 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3fa19dbc-a10c-4863-bf25-035f9d6579e9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.397493 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3fa19dbc-a10c-4863-bf25-035f9d6579e9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.397530 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3fa19dbc-a10c-4863-bf25-035f9d6579e9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.397564 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3fa19dbc-a10c-4863-bf25-035f9d6579e9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.397581 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3fa19dbc-a10c-4863-bf25-035f9d6579e9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.397604 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3fa19dbc-a10c-4863-bf25-035f9d6579e9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.397674 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3fa19dbc-a10c-4863-bf25-035f9d6579e9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.397744 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3fa19dbc-a10c-4863-bf25-035f9d6579e9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.397830 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.397860 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3fa19dbc-a10c-4863-bf25-035f9d6579e9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.398658 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3fa19dbc-a10c-4863-bf25-035f9d6579e9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.401177 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3fa19dbc-a10c-4863-bf25-035f9d6579e9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.402160 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.402327 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3fa19dbc-a10c-4863-bf25-035f9d6579e9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.402680 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3fa19dbc-a10c-4863-bf25-035f9d6579e9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.403256 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3fa19dbc-a10c-4863-bf25-035f9d6579e9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.405000 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3fa19dbc-a10c-4863-bf25-035f9d6579e9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.405588 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3fa19dbc-a10c-4863-bf25-035f9d6579e9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.407487 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3fa19dbc-a10c-4863-bf25-035f9d6579e9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.422754 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3fa19dbc-a10c-4863-bf25-035f9d6579e9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.424389 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zvxq\" (UniqueName: \"kubernetes.io/projected/3fa19dbc-a10c-4863-bf25-035f9d6579e9-kube-api-access-9zvxq\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.448250 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3fa19dbc-a10c-4863-bf25-035f9d6579e9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.716304 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.803597 4771 scope.go:117] "RemoveContainer" containerID="63aadd758b05567dc1365ba3e674b8ba28ce7f2c61ec9a5fcd81f0d9682d8a7c" Oct 02 10:05:54 crc kubenswrapper[4771]: E1002 10:05:54.811594 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Oct 02 10:05:54 crc kubenswrapper[4771]: E1002 10:05:54.811636 4771 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Oct 02 10:05:54 crc kubenswrapper[4771]: E1002 10:05:54.811790 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qwlxb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-h99xg_openstack(62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:05:54 crc kubenswrapper[4771]: E1002 10:05:54.812957 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-h99xg" podUID="62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.933021 4771 scope.go:117] "RemoveContainer" containerID="82fe7b3800926b8bc7dfa8813efaf7f0dd9c638720dde9673e1bf775cfa6e1dd" Oct 02 10:05:54 crc kubenswrapper[4771]: I1002 10:05:54.991876 4771 scope.go:117] "RemoveContainer" containerID="74e0addf34252d1465b8adca2fc82ba42dcd867d6b03aac7aec4c69a2f14976d" Oct 02 10:05:55 crc kubenswrapper[4771]: I1002 10:05:55.437693 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-vzhpr"] Oct 02 10:05:55 crc kubenswrapper[4771]: I1002 10:05:55.453919 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 10:05:55 crc kubenswrapper[4771]: W1002 10:05:55.460257 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07281b0c_ff0a_4195_9ddb_521645890aee.slice/crio-acb228683b73f156f2d326617e45514479b4ed9e0e3670d352dca02afd88670c WatchSource:0}: Error finding container acb228683b73f156f2d326617e45514479b4ed9e0e3670d352dca02afd88670c: Status 404 returned error can't find the container with id acb228683b73f156f2d326617e45514479b4ed9e0e3670d352dca02afd88670c Oct 02 10:05:55 crc kubenswrapper[4771]: I1002 10:05:55.635562 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" event={"ID":"faa0b798-d3ee-4469-b60b-15c77ecf1753","Type":"ContainerStarted","Data":"c57bd4a7a1c6d9ffde2b81658934bd665e3d45945726ed4b2ac340fcc0c44c51"} Oct 02 10:05:55 crc kubenswrapper[4771]: I1002 10:05:55.647090 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"07281b0c-ff0a-4195-9ddb-521645890aee","Type":"ContainerStarted","Data":"acb228683b73f156f2d326617e45514479b4ed9e0e3670d352dca02afd88670c"} Oct 02 10:05:55 crc kubenswrapper[4771]: E1002 10:05:55.649663 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-h99xg" podUID="62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02" Oct 02 10:05:55 crc kubenswrapper[4771]: I1002 10:05:55.679834 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 10:05:55 crc kubenswrapper[4771]: W1002 10:05:55.699696 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fa19dbc_a10c_4863_bf25_035f9d6579e9.slice/crio-042683197f152e6efd9aedcae8cfa33ef2390c1d9f0ae7b7d634bb58539840d1 WatchSource:0}: Error finding container 042683197f152e6efd9aedcae8cfa33ef2390c1d9f0ae7b7d634bb58539840d1: Status 404 returned error can't find the container with id 042683197f152e6efd9aedcae8cfa33ef2390c1d9f0ae7b7d634bb58539840d1 Oct 02 10:05:55 crc kubenswrapper[4771]: I1002 10:05:55.717058 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7dad196-4ebc-43e0-b850-0820f2b072c9" path="/var/lib/kubelet/pods/b7dad196-4ebc-43e0-b850-0820f2b072c9/volumes" Oct 02 10:05:55 crc kubenswrapper[4771]: I1002 10:05:55.718353 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebc075f5-e2ba-4789-a38e-4065fe661ada" path="/var/lib/kubelet/pods/ebc075f5-e2ba-4789-a38e-4065fe661ada/volumes" Oct 02 10:05:56 crc kubenswrapper[4771]: I1002 10:05:56.667709 4771 generic.go:334] "Generic (PLEG): container finished" podID="faa0b798-d3ee-4469-b60b-15c77ecf1753" containerID="0cc2c7e6ce97e816412af6668269194da7865446e929aed22f66c7deeb55abb2" exitCode=0 Oct 02 10:05:56 crc kubenswrapper[4771]: I1002 10:05:56.669488 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" event={"ID":"faa0b798-d3ee-4469-b60b-15c77ecf1753","Type":"ContainerDied","Data":"0cc2c7e6ce97e816412af6668269194da7865446e929aed22f66c7deeb55abb2"} Oct 02 10:05:56 crc kubenswrapper[4771]: I1002 10:05:56.671519 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3fa19dbc-a10c-4863-bf25-035f9d6579e9","Type":"ContainerStarted","Data":"042683197f152e6efd9aedcae8cfa33ef2390c1d9f0ae7b7d634bb58539840d1"} Oct 02 10:05:56 crc kubenswrapper[4771]: I1002 10:05:56.675710 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53db4baf-fd23-4d39-a4c8-0019beae7703","Type":"ContainerStarted","Data":"a41ca7a60b8d874de54ae77b7e93463870e64c38776516de70bf3c7422994816"} Oct 02 10:05:57 crc kubenswrapper[4771]: I1002 10:05:57.687795 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:05:57 crc kubenswrapper[4771]: E1002 10:05:57.689225 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:05:57 crc kubenswrapper[4771]: I1002 10:05:57.702648 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53db4baf-fd23-4d39-a4c8-0019beae7703","Type":"ContainerStarted","Data":"06d6623ab3721e8a30d8e0e813e41cb3f5ca83f5553d33127c0fc352609b546c"} Oct 02 10:05:57 crc kubenswrapper[4771]: I1002 10:05:57.704921 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" event={"ID":"faa0b798-d3ee-4469-b60b-15c77ecf1753","Type":"ContainerStarted","Data":"078065b4b74c08d4791fa8c0d0a801f1b28c6946e4bdbc3f4e3f207251a41de8"} Oct 02 10:05:57 crc kubenswrapper[4771]: I1002 10:05:57.705071 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:05:57 crc kubenswrapper[4771]: I1002 10:05:57.706301 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"07281b0c-ff0a-4195-9ddb-521645890aee","Type":"ContainerStarted","Data":"bf6eb97c981bdff5aab951eb0814694a6f245424dbf8b6c0f7789b003c6396db"} Oct 02 10:05:57 crc kubenswrapper[4771]: I1002 10:05:57.737189 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" podStartSLOduration=9.737167213 podStartE2EDuration="9.737167213s" podCreationTimestamp="2025-10-02 10:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:05:57.725037052 +0000 UTC m=+1745.372722119" watchObservedRunningTime="2025-10-02 10:05:57.737167213 +0000 UTC m=+1745.384852280" Oct 02 10:05:58 crc kubenswrapper[4771]: I1002 10:05:58.721056 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3fa19dbc-a10c-4863-bf25-035f9d6579e9","Type":"ContainerStarted","Data":"3f2cee3926b083e0e3c79c65ccdac96808338a180394cf6e04c15a643ada66e9"} Oct 02 10:05:59 crc kubenswrapper[4771]: E1002 10:05:59.045528 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="53db4baf-fd23-4d39-a4c8-0019beae7703" Oct 02 10:05:59 crc kubenswrapper[4771]: I1002 10:05:59.742708 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53db4baf-fd23-4d39-a4c8-0019beae7703","Type":"ContainerStarted","Data":"bfb61abecb46733d8a3e527ba5a6f7d0d737ce27a7e31250ed39c46ebf30018d"} Oct 02 10:05:59 crc kubenswrapper[4771]: E1002 10:05:59.746237 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="53db4baf-fd23-4d39-a4c8-0019beae7703" Oct 02 10:06:00 crc kubenswrapper[4771]: I1002 10:06:00.779722 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:06:00 crc kubenswrapper[4771]: E1002 10:06:00.783691 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="53db4baf-fd23-4d39-a4c8-0019beae7703" Oct 02 10:06:01 crc kubenswrapper[4771]: E1002 10:06:01.800196 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="53db4baf-fd23-4d39-a4c8-0019beae7703" Oct 02 10:06:03 crc kubenswrapper[4771]: I1002 10:06:03.892320 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f6df4f56c-vzhpr" Oct 02 10:06:03 crc kubenswrapper[4771]: I1002 10:06:03.966213 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-rzgws"] Oct 02 10:06:03 crc kubenswrapper[4771]: I1002 10:06:03.966519 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" podUID="520c351c-db61-4f7b-8c3f-3239b619f553" containerName="dnsmasq-dns" containerID="cri-o://c6e483c524ae041bc3f5309f41b8d5f92d6d95332ef9a6be6377b18fc73abd3d" gracePeriod=10 Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.649388 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.721016 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-ovsdbserver-sb\") pod \"520c351c-db61-4f7b-8c3f-3239b619f553\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.721173 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-config\") pod \"520c351c-db61-4f7b-8c3f-3239b619f553\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.721303 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt5xt\" (UniqueName: \"kubernetes.io/projected/520c351c-db61-4f7b-8c3f-3239b619f553-kube-api-access-gt5xt\") pod \"520c351c-db61-4f7b-8c3f-3239b619f553\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.721445 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-dns-svc\") pod \"520c351c-db61-4f7b-8c3f-3239b619f553\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.721533 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-dns-swift-storage-0\") pod \"520c351c-db61-4f7b-8c3f-3239b619f553\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.721611 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-ovsdbserver-nb\") pod \"520c351c-db61-4f7b-8c3f-3239b619f553\" (UID: \"520c351c-db61-4f7b-8c3f-3239b619f553\") " Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.737983 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/520c351c-db61-4f7b-8c3f-3239b619f553-kube-api-access-gt5xt" (OuterVolumeSpecName: "kube-api-access-gt5xt") pod "520c351c-db61-4f7b-8c3f-3239b619f553" (UID: "520c351c-db61-4f7b-8c3f-3239b619f553"). InnerVolumeSpecName "kube-api-access-gt5xt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.791272 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "520c351c-db61-4f7b-8c3f-3239b619f553" (UID: "520c351c-db61-4f7b-8c3f-3239b619f553"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.796343 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "520c351c-db61-4f7b-8c3f-3239b619f553" (UID: "520c351c-db61-4f7b-8c3f-3239b619f553"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.803269 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-config" (OuterVolumeSpecName: "config") pod "520c351c-db61-4f7b-8c3f-3239b619f553" (UID: "520c351c-db61-4f7b-8c3f-3239b619f553"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.817643 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "520c351c-db61-4f7b-8c3f-3239b619f553" (UID: "520c351c-db61-4f7b-8c3f-3239b619f553"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.826059 4771 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.826333 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.826446 4771 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.827219 4771 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.827330 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt5xt\" (UniqueName: \"kubernetes.io/projected/520c351c-db61-4f7b-8c3f-3239b619f553-kube-api-access-gt5xt\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.827181 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "520c351c-db61-4f7b-8c3f-3239b619f553" (UID: "520c351c-db61-4f7b-8c3f-3239b619f553"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.835557 4771 generic.go:334] "Generic (PLEG): container finished" podID="520c351c-db61-4f7b-8c3f-3239b619f553" containerID="c6e483c524ae041bc3f5309f41b8d5f92d6d95332ef9a6be6377b18fc73abd3d" exitCode=0 Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.835824 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" event={"ID":"520c351c-db61-4f7b-8c3f-3239b619f553","Type":"ContainerDied","Data":"c6e483c524ae041bc3f5309f41b8d5f92d6d95332ef9a6be6377b18fc73abd3d"} Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.835948 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" event={"ID":"520c351c-db61-4f7b-8c3f-3239b619f553","Type":"ContainerDied","Data":"7c22ad2044a31aaf13264b56d941c30d5a638f6ae0a8352ee37dc64641055cad"} Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.836050 4771 scope.go:117] "RemoveContainer" containerID="c6e483c524ae041bc3f5309f41b8d5f92d6d95332ef9a6be6377b18fc73abd3d" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.836346 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-rzgws" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.905180 4771 scope.go:117] "RemoveContainer" containerID="652826d725efd479f42346177f86c5e5979d27f61ea69ff16ef1062e8214140d" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.920459 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-rzgws"] Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.929809 4771 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/520c351c-db61-4f7b-8c3f-3239b619f553-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.935239 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-rzgws"] Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.940342 4771 scope.go:117] "RemoveContainer" containerID="c6e483c524ae041bc3f5309f41b8d5f92d6d95332ef9a6be6377b18fc73abd3d" Oct 02 10:06:04 crc kubenswrapper[4771]: E1002 10:06:04.940717 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6e483c524ae041bc3f5309f41b8d5f92d6d95332ef9a6be6377b18fc73abd3d\": container with ID starting with c6e483c524ae041bc3f5309f41b8d5f92d6d95332ef9a6be6377b18fc73abd3d not found: ID does not exist" containerID="c6e483c524ae041bc3f5309f41b8d5f92d6d95332ef9a6be6377b18fc73abd3d" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.940842 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6e483c524ae041bc3f5309f41b8d5f92d6d95332ef9a6be6377b18fc73abd3d"} err="failed to get container status \"c6e483c524ae041bc3f5309f41b8d5f92d6d95332ef9a6be6377b18fc73abd3d\": rpc error: code = NotFound desc = could not find container \"c6e483c524ae041bc3f5309f41b8d5f92d6d95332ef9a6be6377b18fc73abd3d\": container with ID starting with c6e483c524ae041bc3f5309f41b8d5f92d6d95332ef9a6be6377b18fc73abd3d not found: ID does not exist" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.940987 4771 scope.go:117] "RemoveContainer" containerID="652826d725efd479f42346177f86c5e5979d27f61ea69ff16ef1062e8214140d" Oct 02 10:06:04 crc kubenswrapper[4771]: E1002 10:06:04.941846 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"652826d725efd479f42346177f86c5e5979d27f61ea69ff16ef1062e8214140d\": container with ID starting with 652826d725efd479f42346177f86c5e5979d27f61ea69ff16ef1062e8214140d not found: ID does not exist" containerID="652826d725efd479f42346177f86c5e5979d27f61ea69ff16ef1062e8214140d" Oct 02 10:06:04 crc kubenswrapper[4771]: I1002 10:06:04.941884 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"652826d725efd479f42346177f86c5e5979d27f61ea69ff16ef1062e8214140d"} err="failed to get container status \"652826d725efd479f42346177f86c5e5979d27f61ea69ff16ef1062e8214140d\": rpc error: code = NotFound desc = could not find container \"652826d725efd479f42346177f86c5e5979d27f61ea69ff16ef1062e8214140d\": container with ID starting with 652826d725efd479f42346177f86c5e5979d27f61ea69ff16ef1062e8214140d not found: ID does not exist" Oct 02 10:06:05 crc kubenswrapper[4771]: I1002 10:06:05.705823 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="520c351c-db61-4f7b-8c3f-3239b619f553" path="/var/lib/kubelet/pods/520c351c-db61-4f7b-8c3f-3239b619f553/volumes" Oct 02 10:06:08 crc kubenswrapper[4771]: I1002 10:06:08.681933 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:06:08 crc kubenswrapper[4771]: E1002 10:06:08.682804 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:06:08 crc kubenswrapper[4771]: I1002 10:06:08.892520 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-h99xg" event={"ID":"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02","Type":"ContainerStarted","Data":"90023531a6d5319f53f2bc581196dbbda23b992e70a511f264d7fec38a75689a"} Oct 02 10:06:08 crc kubenswrapper[4771]: I1002 10:06:08.921248 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-h99xg" podStartSLOduration=2.168896577 podStartE2EDuration="42.921216842s" podCreationTimestamp="2025-10-02 10:05:26 +0000 UTC" firstStartedPulling="2025-10-02 10:05:27.132285952 +0000 UTC m=+1714.779971019" lastFinishedPulling="2025-10-02 10:06:07.884606217 +0000 UTC m=+1755.532291284" observedRunningTime="2025-10-02 10:06:08.912116511 +0000 UTC m=+1756.559801578" watchObservedRunningTime="2025-10-02 10:06:08.921216842 +0000 UTC m=+1756.568901919" Oct 02 10:06:10 crc kubenswrapper[4771]: I1002 10:06:10.931328 4771 generic.go:334] "Generic (PLEG): container finished" podID="62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02" containerID="90023531a6d5319f53f2bc581196dbbda23b992e70a511f264d7fec38a75689a" exitCode=0 Oct 02 10:06:10 crc kubenswrapper[4771]: I1002 10:06:10.931537 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-h99xg" event={"ID":"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02","Type":"ContainerDied","Data":"90023531a6d5319f53f2bc581196dbbda23b992e70a511f264d7fec38a75689a"} Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.580441 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-h99xg" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.635814 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-config-data\") pod \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\" (UID: \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\") " Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.636883 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwlxb\" (UniqueName: \"kubernetes.io/projected/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-kube-api-access-qwlxb\") pod \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\" (UID: \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\") " Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.637113 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-combined-ca-bundle\") pod \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\" (UID: \"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02\") " Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.647662 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-kube-api-access-qwlxb" (OuterVolumeSpecName: "kube-api-access-qwlxb") pod "62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02" (UID: "62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02"). InnerVolumeSpecName "kube-api-access-qwlxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.687020 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02" (UID: "62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.746238 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwlxb\" (UniqueName: \"kubernetes.io/projected/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-kube-api-access-qwlxb\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.746279 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.774330 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-config-data" (OuterVolumeSpecName: "config-data") pod "62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02" (UID: "62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.848750 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.936908 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj"] Oct 02 10:06:12 crc kubenswrapper[4771]: E1002 10:06:12.937605 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="520c351c-db61-4f7b-8c3f-3239b619f553" containerName="init" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.937631 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="520c351c-db61-4f7b-8c3f-3239b619f553" containerName="init" Oct 02 10:06:12 crc kubenswrapper[4771]: E1002 10:06:12.937673 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02" containerName="heat-db-sync" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.937681 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02" containerName="heat-db-sync" Oct 02 10:06:12 crc kubenswrapper[4771]: E1002 10:06:12.937704 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="520c351c-db61-4f7b-8c3f-3239b619f553" containerName="dnsmasq-dns" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.937712 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="520c351c-db61-4f7b-8c3f-3239b619f553" containerName="dnsmasq-dns" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.938032 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="520c351c-db61-4f7b-8c3f-3239b619f553" containerName="dnsmasq-dns" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.938063 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02" containerName="heat-db-sync" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.939175 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.942342 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.942589 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.942909 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.947700 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.973525 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj"] Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.977753 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-h99xg" event={"ID":"62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02","Type":"ContainerDied","Data":"a9ecbee9db07b00f02d9a94847eaebe2b1a470b1f8488350a133ae07b92bf51b"} Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.977808 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9ecbee9db07b00f02d9a94847eaebe2b1a470b1f8488350a133ae07b92bf51b" Oct 02 10:06:12 crc kubenswrapper[4771]: I1002 10:06:12.977881 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-h99xg" Oct 02 10:06:13 crc kubenswrapper[4771]: I1002 10:06:13.053797 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:13 crc kubenswrapper[4771]: I1002 10:06:13.054306 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:13 crc kubenswrapper[4771]: I1002 10:06:13.054552 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:13 crc kubenswrapper[4771]: I1002 10:06:13.054633 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj8k4\" (UniqueName: \"kubernetes.io/projected/58b8222a-a342-4b38-8aa6-47c6ea065493-kube-api-access-lj8k4\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:13 crc kubenswrapper[4771]: I1002 10:06:13.157723 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:13 crc kubenswrapper[4771]: I1002 10:06:13.157852 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:13 crc kubenswrapper[4771]: I1002 10:06:13.157911 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:13 crc kubenswrapper[4771]: I1002 10:06:13.157946 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj8k4\" (UniqueName: \"kubernetes.io/projected/58b8222a-a342-4b38-8aa6-47c6ea065493-kube-api-access-lj8k4\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:13 crc kubenswrapper[4771]: I1002 10:06:13.163757 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:13 crc kubenswrapper[4771]: I1002 10:06:13.165532 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:13 crc kubenswrapper[4771]: I1002 10:06:13.166704 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:13 crc kubenswrapper[4771]: I1002 10:06:13.179325 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj8k4\" (UniqueName: \"kubernetes.io/projected/58b8222a-a342-4b38-8aa6-47c6ea065493-kube-api-access-lj8k4\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:13 crc kubenswrapper[4771]: I1002 10:06:13.272278 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.088470 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-54b646f5dd-mdqrv"] Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.090630 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.103868 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-54b646f5dd-mdqrv"] Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.195899 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-58b9db67cd-ctlwn"] Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.202384 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f645324-58c1-45d2-9859-aa05f9644975-config-data-custom\") pod \"heat-engine-54b646f5dd-mdqrv\" (UID: \"2f645324-58c1-45d2-9859-aa05f9644975\") " pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.202634 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f645324-58c1-45d2-9859-aa05f9644975-combined-ca-bundle\") pod \"heat-engine-54b646f5dd-mdqrv\" (UID: \"2f645324-58c1-45d2-9859-aa05f9644975\") " pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.202695 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f645324-58c1-45d2-9859-aa05f9644975-config-data\") pod \"heat-engine-54b646f5dd-mdqrv\" (UID: \"2f645324-58c1-45d2-9859-aa05f9644975\") " pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.202886 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5mtg\" (UniqueName: \"kubernetes.io/projected/2f645324-58c1-45d2-9859-aa05f9644975-kube-api-access-d5mtg\") pod \"heat-engine-54b646f5dd-mdqrv\" (UID: \"2f645324-58c1-45d2-9859-aa05f9644975\") " pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.203259 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.243356 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-f9d48f8dd-cs4p5"] Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.246163 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.291862 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-f9d48f8dd-cs4p5"] Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306140 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f645324-58c1-45d2-9859-aa05f9644975-combined-ca-bundle\") pod \"heat-engine-54b646f5dd-mdqrv\" (UID: \"2f645324-58c1-45d2-9859-aa05f9644975\") " pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306202 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f645324-58c1-45d2-9859-aa05f9644975-config-data\") pod \"heat-engine-54b646f5dd-mdqrv\" (UID: \"2f645324-58c1-45d2-9859-aa05f9644975\") " pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306282 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-internal-tls-certs\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306348 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-config-data-custom\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306402 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5mtg\" (UniqueName: \"kubernetes.io/projected/2f645324-58c1-45d2-9859-aa05f9644975-kube-api-access-d5mtg\") pod \"heat-engine-54b646f5dd-mdqrv\" (UID: \"2f645324-58c1-45d2-9859-aa05f9644975\") " pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306425 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-combined-ca-bundle\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306451 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-combined-ca-bundle\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306516 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7bwf\" (UniqueName: \"kubernetes.io/projected/3350ec2f-e01d-437e-99fd-49a97abfd448-kube-api-access-w7bwf\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306539 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-config-data\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306576 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-config-data-custom\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306632 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f645324-58c1-45d2-9859-aa05f9644975-config-data-custom\") pod \"heat-engine-54b646f5dd-mdqrv\" (UID: \"2f645324-58c1-45d2-9859-aa05f9644975\") " pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306671 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9phlp\" (UniqueName: \"kubernetes.io/projected/d8ca2772-c933-4f37-9f65-65ff09edfc8d-kube-api-access-9phlp\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306700 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-internal-tls-certs\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306719 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-public-tls-certs\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306756 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-config-data\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.306777 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-public-tls-certs\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.319723 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f645324-58c1-45d2-9859-aa05f9644975-config-data-custom\") pod \"heat-engine-54b646f5dd-mdqrv\" (UID: \"2f645324-58c1-45d2-9859-aa05f9644975\") " pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.320249 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f645324-58c1-45d2-9859-aa05f9644975-config-data\") pod \"heat-engine-54b646f5dd-mdqrv\" (UID: \"2f645324-58c1-45d2-9859-aa05f9644975\") " pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.325306 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f645324-58c1-45d2-9859-aa05f9644975-combined-ca-bundle\") pod \"heat-engine-54b646f5dd-mdqrv\" (UID: \"2f645324-58c1-45d2-9859-aa05f9644975\") " pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.331746 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-58b9db67cd-ctlwn"] Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.332260 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5mtg\" (UniqueName: \"kubernetes.io/projected/2f645324-58c1-45d2-9859-aa05f9644975-kube-api-access-d5mtg\") pod \"heat-engine-54b646f5dd-mdqrv\" (UID: \"2f645324-58c1-45d2-9859-aa05f9644975\") " pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.412790 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-internal-tls-certs\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.412891 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-config-data-custom\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.412947 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-combined-ca-bundle\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.412978 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-combined-ca-bundle\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.413049 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7bwf\" (UniqueName: \"kubernetes.io/projected/3350ec2f-e01d-437e-99fd-49a97abfd448-kube-api-access-w7bwf\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.413079 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-config-data\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.413121 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-config-data-custom\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.413196 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9phlp\" (UniqueName: \"kubernetes.io/projected/d8ca2772-c933-4f37-9f65-65ff09edfc8d-kube-api-access-9phlp\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.413227 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-internal-tls-certs\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.413252 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-public-tls-certs\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.413297 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-config-data\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.413324 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-public-tls-certs\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.433193 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-config-data-custom\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.434242 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-config-data-custom\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.441741 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.443200 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-config-data\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.443708 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-internal-tls-certs\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.444607 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-combined-ca-bundle\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.447175 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7bwf\" (UniqueName: \"kubernetes.io/projected/3350ec2f-e01d-437e-99fd-49a97abfd448-kube-api-access-w7bwf\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.466585 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-config-data\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.471633 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-internal-tls-certs\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.474321 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9phlp\" (UniqueName: \"kubernetes.io/projected/d8ca2772-c933-4f37-9f65-65ff09edfc8d-kube-api-access-9phlp\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.474366 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-public-tls-certs\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.482386 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8ca2772-c933-4f37-9f65-65ff09edfc8d-public-tls-certs\") pod \"heat-cfnapi-58b9db67cd-ctlwn\" (UID: \"d8ca2772-c933-4f37-9f65-65ff09edfc8d\") " pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.497620 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3350ec2f-e01d-437e-99fd-49a97abfd448-combined-ca-bundle\") pod \"heat-api-f9d48f8dd-cs4p5\" (UID: \"3350ec2f-e01d-437e-99fd-49a97abfd448\") " pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.549662 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.576941 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.722790 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 10:06:14 crc kubenswrapper[4771]: I1002 10:06:14.827006 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj"] Oct 02 10:06:15 crc kubenswrapper[4771]: I1002 10:06:15.037616 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" event={"ID":"58b8222a-a342-4b38-8aa6-47c6ea065493","Type":"ContainerStarted","Data":"d5a13739211c2f5ff92c31dd5cc4af442483456044c2a1ee291e53f21f800e35"} Oct 02 10:06:15 crc kubenswrapper[4771]: I1002 10:06:15.113563 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-54b646f5dd-mdqrv"] Oct 02 10:06:15 crc kubenswrapper[4771]: I1002 10:06:15.373726 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-f9d48f8dd-cs4p5"] Oct 02 10:06:15 crc kubenswrapper[4771]: I1002 10:06:15.589879 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-58b9db67cd-ctlwn"] Oct 02 10:06:15 crc kubenswrapper[4771]: W1002 10:06:15.593885 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8ca2772_c933_4f37_9f65_65ff09edfc8d.slice/crio-0dbd341c9a32e5aa4f4d7f30cf95bf3f2584a147035a5bee5bf964a4226c7b40 WatchSource:0}: Error finding container 0dbd341c9a32e5aa4f4d7f30cf95bf3f2584a147035a5bee5bf964a4226c7b40: Status 404 returned error can't find the container with id 0dbd341c9a32e5aa4f4d7f30cf95bf3f2584a147035a5bee5bf964a4226c7b40 Oct 02 10:06:16 crc kubenswrapper[4771]: I1002 10:06:16.066367 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-54b646f5dd-mdqrv" event={"ID":"2f645324-58c1-45d2-9859-aa05f9644975","Type":"ContainerStarted","Data":"a851c8b1d71a2fa70afd99ab49bfe119132df2532073fe672ff1d19f6649589f"} Oct 02 10:06:16 crc kubenswrapper[4771]: I1002 10:06:16.066415 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-54b646f5dd-mdqrv" event={"ID":"2f645324-58c1-45d2-9859-aa05f9644975","Type":"ContainerStarted","Data":"c838ec00ef41e3cc72d62f6c0739c0a76c0a9e61092a5b9cdfa65ad067594c9b"} Oct 02 10:06:16 crc kubenswrapper[4771]: I1002 10:06:16.066494 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:16 crc kubenswrapper[4771]: I1002 10:06:16.079852 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53db4baf-fd23-4d39-a4c8-0019beae7703","Type":"ContainerStarted","Data":"dfddb7195a3372f2521ccce92b278231d8ea60ab51d8d2f205d0f24d34b44756"} Oct 02 10:06:16 crc kubenswrapper[4771]: I1002 10:06:16.083649 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f9d48f8dd-cs4p5" event={"ID":"3350ec2f-e01d-437e-99fd-49a97abfd448","Type":"ContainerStarted","Data":"26b1d4b36e2e40f17095ad7ee8e3166ab630d8ef92887ccb6e5c7b80b751eb5b"} Oct 02 10:06:16 crc kubenswrapper[4771]: I1002 10:06:16.086835 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" event={"ID":"d8ca2772-c933-4f37-9f65-65ff09edfc8d","Type":"ContainerStarted","Data":"0dbd341c9a32e5aa4f4d7f30cf95bf3f2584a147035a5bee5bf964a4226c7b40"} Oct 02 10:06:16 crc kubenswrapper[4771]: I1002 10:06:16.097605 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-54b646f5dd-mdqrv" podStartSLOduration=2.097586832 podStartE2EDuration="2.097586832s" podCreationTimestamp="2025-10-02 10:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:06:16.092812846 +0000 UTC m=+1763.740497913" watchObservedRunningTime="2025-10-02 10:06:16.097586832 +0000 UTC m=+1763.745271899" Oct 02 10:06:16 crc kubenswrapper[4771]: I1002 10:06:16.130188 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.487769675 podStartE2EDuration="45.130166024s" podCreationTimestamp="2025-10-02 10:05:31 +0000 UTC" firstStartedPulling="2025-10-02 10:05:32.361081567 +0000 UTC m=+1720.008766634" lastFinishedPulling="2025-10-02 10:06:15.003477926 +0000 UTC m=+1762.651162983" observedRunningTime="2025-10-02 10:06:16.118711221 +0000 UTC m=+1763.766396288" watchObservedRunningTime="2025-10-02 10:06:16.130166024 +0000 UTC m=+1763.777851091" Oct 02 10:06:18 crc kubenswrapper[4771]: I1002 10:06:18.127465 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" event={"ID":"d8ca2772-c933-4f37-9f65-65ff09edfc8d","Type":"ContainerStarted","Data":"1726f2f41c8e98416664766dcf59f6ac4cf6f864bf880d1dc0cbb2f37f4960a3"} Oct 02 10:06:18 crc kubenswrapper[4771]: I1002 10:06:18.128141 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:18 crc kubenswrapper[4771]: I1002 10:06:18.130045 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f9d48f8dd-cs4p5" event={"ID":"3350ec2f-e01d-437e-99fd-49a97abfd448","Type":"ContainerStarted","Data":"5225f84900c0b6af44602d51cc284e7e4988b18240bceb676d8e22b4e246565c"} Oct 02 10:06:18 crc kubenswrapper[4771]: I1002 10:06:18.130173 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:18 crc kubenswrapper[4771]: I1002 10:06:18.155189 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" podStartSLOduration=2.022209247 podStartE2EDuration="4.155167329s" podCreationTimestamp="2025-10-02 10:06:14 +0000 UTC" firstStartedPulling="2025-10-02 10:06:15.597659741 +0000 UTC m=+1763.245344808" lastFinishedPulling="2025-10-02 10:06:17.730617823 +0000 UTC m=+1765.378302890" observedRunningTime="2025-10-02 10:06:18.146022307 +0000 UTC m=+1765.793707374" watchObservedRunningTime="2025-10-02 10:06:18.155167329 +0000 UTC m=+1765.802852396" Oct 02 10:06:18 crc kubenswrapper[4771]: I1002 10:06:18.186411 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-f9d48f8dd-cs4p5" podStartSLOduration=1.842653584 podStartE2EDuration="4.186385895s" podCreationTimestamp="2025-10-02 10:06:14 +0000 UTC" firstStartedPulling="2025-10-02 10:06:15.383519693 +0000 UTC m=+1763.031204750" lastFinishedPulling="2025-10-02 10:06:17.727251994 +0000 UTC m=+1765.374937061" observedRunningTime="2025-10-02 10:06:18.17071651 +0000 UTC m=+1765.818401567" watchObservedRunningTime="2025-10-02 10:06:18.186385895 +0000 UTC m=+1765.834070962" Oct 02 10:06:22 crc kubenswrapper[4771]: I1002 10:06:22.681677 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:06:22 crc kubenswrapper[4771]: E1002 10:06:22.682604 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:06:28 crc kubenswrapper[4771]: I1002 10:06:28.464917 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-58b9db67cd-ctlwn" Oct 02 10:06:28 crc kubenswrapper[4771]: I1002 10:06:28.474327 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-f9d48f8dd-cs4p5" Oct 02 10:06:28 crc kubenswrapper[4771]: I1002 10:06:28.623802 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5cc7b668b9-5jkkb"] Oct 02 10:06:28 crc kubenswrapper[4771]: I1002 10:06:28.624534 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" podUID="f740f4c4-0587-4077-8a4c-3135aea31447" containerName="heat-cfnapi" containerID="cri-o://05e51e74039fab8e39bc3c070ba2da03f715ddff25ed226df376e4cc45e96f3f" gracePeriod=60 Oct 02 10:06:28 crc kubenswrapper[4771]: I1002 10:06:28.652449 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6dfdb97779-bx9w5"] Oct 02 10:06:28 crc kubenswrapper[4771]: I1002 10:06:28.652752 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-6dfdb97779-bx9w5" podUID="ff726ea6-69e3-4d3e-98b5-099265579c80" containerName="heat-api" containerID="cri-o://aaa1d2cffcab599d96ef4dc30e9419a93554d739b80468d87c9266d4999d5f94" gracePeriod=60 Oct 02 10:06:29 crc kubenswrapper[4771]: I1002 10:06:29.268504 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" event={"ID":"58b8222a-a342-4b38-8aa6-47c6ea065493","Type":"ContainerStarted","Data":"a8b57ff6bd1d5ac75681e27a32751615b1abd13286e27c4275e67f3c8692579a"} Oct 02 10:06:29 crc kubenswrapper[4771]: I1002 10:06:29.300260 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" podStartSLOduration=3.716549349 podStartE2EDuration="17.300236196s" podCreationTimestamp="2025-10-02 10:06:12 +0000 UTC" firstStartedPulling="2025-10-02 10:06:14.834076613 +0000 UTC m=+1762.481761680" lastFinishedPulling="2025-10-02 10:06:28.41776346 +0000 UTC m=+1776.065448527" observedRunningTime="2025-10-02 10:06:29.285047174 +0000 UTC m=+1776.932732241" watchObservedRunningTime="2025-10-02 10:06:29.300236196 +0000 UTC m=+1776.947921263" Oct 02 10:06:30 crc kubenswrapper[4771]: I1002 10:06:30.284292 4771 generic.go:334] "Generic (PLEG): container finished" podID="3fa19dbc-a10c-4863-bf25-035f9d6579e9" containerID="3f2cee3926b083e0e3c79c65ccdac96808338a180394cf6e04c15a643ada66e9" exitCode=0 Oct 02 10:06:30 crc kubenswrapper[4771]: I1002 10:06:30.284395 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3fa19dbc-a10c-4863-bf25-035f9d6579e9","Type":"ContainerDied","Data":"3f2cee3926b083e0e3c79c65ccdac96808338a180394cf6e04c15a643ada66e9"} Oct 02 10:06:30 crc kubenswrapper[4771]: I1002 10:06:30.290944 4771 generic.go:334] "Generic (PLEG): container finished" podID="07281b0c-ff0a-4195-9ddb-521645890aee" containerID="bf6eb97c981bdff5aab951eb0814694a6f245424dbf8b6c0f7789b003c6396db" exitCode=0 Oct 02 10:06:30 crc kubenswrapper[4771]: I1002 10:06:30.291545 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"07281b0c-ff0a-4195-9ddb-521645890aee","Type":"ContainerDied","Data":"bf6eb97c981bdff5aab951eb0814694a6f245424dbf8b6c0f7789b003c6396db"} Oct 02 10:06:31 crc kubenswrapper[4771]: I1002 10:06:31.304150 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3fa19dbc-a10c-4863-bf25-035f9d6579e9","Type":"ContainerStarted","Data":"8a4315097ae06dce36290cb0e34baf88b98c5d8b715a701a5a98e9a9bb6741c8"} Oct 02 10:06:31 crc kubenswrapper[4771]: I1002 10:06:31.305944 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:06:31 crc kubenswrapper[4771]: I1002 10:06:31.306545 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"07281b0c-ff0a-4195-9ddb-521645890aee","Type":"ContainerStarted","Data":"cd62e02916050d72daaf871d4c32d9f35b60480076f25d28c40866ddbcc2c9e1"} Oct 02 10:06:31 crc kubenswrapper[4771]: I1002 10:06:31.306762 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 10:06:31 crc kubenswrapper[4771]: I1002 10:06:31.332886 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.332865381 podStartE2EDuration="38.332865381s" podCreationTimestamp="2025-10-02 10:05:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:06:31.327214832 +0000 UTC m=+1778.974899899" watchObservedRunningTime="2025-10-02 10:06:31.332865381 +0000 UTC m=+1778.980550448" Oct 02 10:06:31 crc kubenswrapper[4771]: I1002 10:06:31.381979 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.381955701 podStartE2EDuration="38.381955701s" podCreationTimestamp="2025-10-02 10:05:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:06:31.374545764 +0000 UTC m=+1779.022230841" watchObservedRunningTime="2025-10-02 10:06:31.381955701 +0000 UTC m=+1779.029640778" Oct 02 10:06:31 crc kubenswrapper[4771]: I1002 10:06:31.774140 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" podUID="f740f4c4-0587-4077-8a4c-3135aea31447" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.223:8000/healthcheck\": read tcp 10.217.0.2:58362->10.217.0.223:8000: read: connection reset by peer" Oct 02 10:06:31 crc kubenswrapper[4771]: I1002 10:06:31.809797 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-6dfdb97779-bx9w5" podUID="ff726ea6-69e3-4d3e-98b5-099265579c80" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.222:8004/healthcheck\": read tcp 10.217.0.2:38544->10.217.0.222:8004: read: connection reset by peer" Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.334444 4771 generic.go:334] "Generic (PLEG): container finished" podID="f740f4c4-0587-4077-8a4c-3135aea31447" containerID="05e51e74039fab8e39bc3c070ba2da03f715ddff25ed226df376e4cc45e96f3f" exitCode=0 Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.334673 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" event={"ID":"f740f4c4-0587-4077-8a4c-3135aea31447","Type":"ContainerDied","Data":"05e51e74039fab8e39bc3c070ba2da03f715ddff25ed226df376e4cc45e96f3f"} Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.340105 4771 generic.go:334] "Generic (PLEG): container finished" podID="ff726ea6-69e3-4d3e-98b5-099265579c80" containerID="aaa1d2cffcab599d96ef4dc30e9419a93554d739b80468d87c9266d4999d5f94" exitCode=0 Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.340291 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6dfdb97779-bx9w5" event={"ID":"ff726ea6-69e3-4d3e-98b5-099265579c80","Type":"ContainerDied","Data":"aaa1d2cffcab599d96ef4dc30e9419a93554d739b80468d87c9266d4999d5f94"} Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.842639 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.867644 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.907164 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-config-data\") pod \"ff726ea6-69e3-4d3e-98b5-099265579c80\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.907279 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vc74\" (UniqueName: \"kubernetes.io/projected/ff726ea6-69e3-4d3e-98b5-099265579c80-kube-api-access-5vc74\") pod \"ff726ea6-69e3-4d3e-98b5-099265579c80\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.907338 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-combined-ca-bundle\") pod \"ff726ea6-69e3-4d3e-98b5-099265579c80\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.907478 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-public-tls-certs\") pod \"ff726ea6-69e3-4d3e-98b5-099265579c80\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.907533 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-config-data-custom\") pod \"ff726ea6-69e3-4d3e-98b5-099265579c80\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.907601 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-internal-tls-certs\") pod \"ff726ea6-69e3-4d3e-98b5-099265579c80\" (UID: \"ff726ea6-69e3-4d3e-98b5-099265579c80\") " Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.918895 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ff726ea6-69e3-4d3e-98b5-099265579c80" (UID: "ff726ea6-69e3-4d3e-98b5-099265579c80"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.926724 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff726ea6-69e3-4d3e-98b5-099265579c80-kube-api-access-5vc74" (OuterVolumeSpecName: "kube-api-access-5vc74") pod "ff726ea6-69e3-4d3e-98b5-099265579c80" (UID: "ff726ea6-69e3-4d3e-98b5-099265579c80"). InnerVolumeSpecName "kube-api-access-5vc74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.991092 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ff726ea6-69e3-4d3e-98b5-099265579c80" (UID: "ff726ea6-69e3-4d3e-98b5-099265579c80"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:32 crc kubenswrapper[4771]: I1002 10:06:32.992355 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff726ea6-69e3-4d3e-98b5-099265579c80" (UID: "ff726ea6-69e3-4d3e-98b5-099265579c80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.009613 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4t5c\" (UniqueName: \"kubernetes.io/projected/f740f4c4-0587-4077-8a4c-3135aea31447-kube-api-access-d4t5c\") pod \"f740f4c4-0587-4077-8a4c-3135aea31447\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.009673 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-config-data-custom\") pod \"f740f4c4-0587-4077-8a4c-3135aea31447\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.009800 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-combined-ca-bundle\") pod \"f740f4c4-0587-4077-8a4c-3135aea31447\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.009886 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-internal-tls-certs\") pod \"f740f4c4-0587-4077-8a4c-3135aea31447\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.009915 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-public-tls-certs\") pod \"f740f4c4-0587-4077-8a4c-3135aea31447\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.009943 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-config-data\") pod \"f740f4c4-0587-4077-8a4c-3135aea31447\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.011166 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vc74\" (UniqueName: \"kubernetes.io/projected/ff726ea6-69e3-4d3e-98b5-099265579c80-kube-api-access-5vc74\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.011187 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.011197 4771 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.011206 4771 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.018418 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f740f4c4-0587-4077-8a4c-3135aea31447" (UID: "f740f4c4-0587-4077-8a4c-3135aea31447"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.018700 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f740f4c4-0587-4077-8a4c-3135aea31447-kube-api-access-d4t5c" (OuterVolumeSpecName: "kube-api-access-d4t5c") pod "f740f4c4-0587-4077-8a4c-3135aea31447" (UID: "f740f4c4-0587-4077-8a4c-3135aea31447"). InnerVolumeSpecName "kube-api-access-d4t5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.032852 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ff726ea6-69e3-4d3e-98b5-099265579c80" (UID: "ff726ea6-69e3-4d3e-98b5-099265579c80"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.070473 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-config-data" (OuterVolumeSpecName: "config-data") pod "ff726ea6-69e3-4d3e-98b5-099265579c80" (UID: "ff726ea6-69e3-4d3e-98b5-099265579c80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.089764 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f740f4c4-0587-4077-8a4c-3135aea31447" (UID: "f740f4c4-0587-4077-8a4c-3135aea31447"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.111653 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f740f4c4-0587-4077-8a4c-3135aea31447" (UID: "f740f4c4-0587-4077-8a4c-3135aea31447"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.113761 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-public-tls-certs\") pod \"f740f4c4-0587-4077-8a4c-3135aea31447\" (UID: \"f740f4c4-0587-4077-8a4c-3135aea31447\") " Oct 02 10:06:33 crc kubenswrapper[4771]: W1002 10:06:33.114315 4771 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/f740f4c4-0587-4077-8a4c-3135aea31447/volumes/kubernetes.io~secret/public-tls-certs Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.114339 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f740f4c4-0587-4077-8a4c-3135aea31447" (UID: "f740f4c4-0587-4077-8a4c-3135aea31447"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.115347 4771 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.115486 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.115567 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4t5c\" (UniqueName: \"kubernetes.io/projected/f740f4c4-0587-4077-8a4c-3135aea31447-kube-api-access-d4t5c\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.115661 4771 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.115738 4771 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff726ea6-69e3-4d3e-98b5-099265579c80-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.115812 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.123384 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f740f4c4-0587-4077-8a4c-3135aea31447" (UID: "f740f4c4-0587-4077-8a4c-3135aea31447"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.137334 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-config-data" (OuterVolumeSpecName: "config-data") pod "f740f4c4-0587-4077-8a4c-3135aea31447" (UID: "f740f4c4-0587-4077-8a4c-3135aea31447"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.219355 4771 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.219399 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f740f4c4-0587-4077-8a4c-3135aea31447-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.358554 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" event={"ID":"f740f4c4-0587-4077-8a4c-3135aea31447","Type":"ContainerDied","Data":"1d6a04c992549b1ed393df3a48c5dbc95049e26272552539b71078e6a6fe2f86"} Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.358584 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5cc7b668b9-5jkkb" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.358641 4771 scope.go:117] "RemoveContainer" containerID="05e51e74039fab8e39bc3c070ba2da03f715ddff25ed226df376e4cc45e96f3f" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.365280 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6dfdb97779-bx9w5" event={"ID":"ff726ea6-69e3-4d3e-98b5-099265579c80","Type":"ContainerDied","Data":"f05783dcc9d8e306cabb96eb5e48ec0cb5208ef7f0f23366fbe3b5a63afa9ab6"} Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.365337 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6dfdb97779-bx9w5" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.398442 4771 scope.go:117] "RemoveContainer" containerID="aaa1d2cffcab599d96ef4dc30e9419a93554d739b80468d87c9266d4999d5f94" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.414829 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5cc7b668b9-5jkkb"] Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.426463 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-5cc7b668b9-5jkkb"] Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.437662 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6dfdb97779-bx9w5"] Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.450705 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-6dfdb97779-bx9w5"] Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.701083 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f740f4c4-0587-4077-8a4c-3135aea31447" path="/var/lib/kubelet/pods/f740f4c4-0587-4077-8a4c-3135aea31447/volumes" Oct 02 10:06:33 crc kubenswrapper[4771]: I1002 10:06:33.701800 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff726ea6-69e3-4d3e-98b5-099265579c80" path="/var/lib/kubelet/pods/ff726ea6-69e3-4d3e-98b5-099265579c80/volumes" Oct 02 10:06:34 crc kubenswrapper[4771]: I1002 10:06:34.497504 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-54b646f5dd-mdqrv" Oct 02 10:06:34 crc kubenswrapper[4771]: I1002 10:06:34.549054 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-868b7bfb89-mjrmr"] Oct 02 10:06:34 crc kubenswrapper[4771]: I1002 10:06:34.549346 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-868b7bfb89-mjrmr" podUID="a4bc89da-2332-412a-b7af-6056f830310a" containerName="heat-engine" containerID="cri-o://763b93cb9e0bd45f8673736c7da2c1d7ebc5601f38a1b419d4f03d0dea0ff226" gracePeriod=60 Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.345036 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-6mtn4"] Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.359782 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-6mtn4"] Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.412525 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-l5pd4"] Oct 02 10:06:36 crc kubenswrapper[4771]: E1002 10:06:36.413377 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f740f4c4-0587-4077-8a4c-3135aea31447" containerName="heat-cfnapi" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.413466 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f740f4c4-0587-4077-8a4c-3135aea31447" containerName="heat-cfnapi" Oct 02 10:06:36 crc kubenswrapper[4771]: E1002 10:06:36.413584 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff726ea6-69e3-4d3e-98b5-099265579c80" containerName="heat-api" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.413643 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff726ea6-69e3-4d3e-98b5-099265579c80" containerName="heat-api" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.413947 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff726ea6-69e3-4d3e-98b5-099265579c80" containerName="heat-api" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.414030 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="f740f4c4-0587-4077-8a4c-3135aea31447" containerName="heat-cfnapi" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.415085 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.426231 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-l5pd4"] Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.507960 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-config-data\") pod \"aodh-db-sync-l5pd4\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.508010 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-combined-ca-bundle\") pod \"aodh-db-sync-l5pd4\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.508037 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm8q7\" (UniqueName: \"kubernetes.io/projected/94e19f78-6425-4099-8188-8dabe4c2d9f0-kube-api-access-wm8q7\") pod \"aodh-db-sync-l5pd4\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.508335 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-scripts\") pod \"aodh-db-sync-l5pd4\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.611277 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-scripts\") pod \"aodh-db-sync-l5pd4\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.611614 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-config-data\") pod \"aodh-db-sync-l5pd4\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.611691 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-combined-ca-bundle\") pod \"aodh-db-sync-l5pd4\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.611724 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm8q7\" (UniqueName: \"kubernetes.io/projected/94e19f78-6425-4099-8188-8dabe4c2d9f0-kube-api-access-wm8q7\") pod \"aodh-db-sync-l5pd4\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.619069 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-scripts\") pod \"aodh-db-sync-l5pd4\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.619654 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-config-data\") pod \"aodh-db-sync-l5pd4\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.630740 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm8q7\" (UniqueName: \"kubernetes.io/projected/94e19f78-6425-4099-8188-8dabe4c2d9f0-kube-api-access-wm8q7\") pod \"aodh-db-sync-l5pd4\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.657077 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-combined-ca-bundle\") pod \"aodh-db-sync-l5pd4\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.683972 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:06:36 crc kubenswrapper[4771]: E1002 10:06:36.684309 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:06:36 crc kubenswrapper[4771]: I1002 10:06:36.775595 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:37 crc kubenswrapper[4771]: I1002 10:06:37.364909 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-l5pd4"] Oct 02 10:06:37 crc kubenswrapper[4771]: I1002 10:06:37.415584 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-l5pd4" event={"ID":"94e19f78-6425-4099-8188-8dabe4c2d9f0","Type":"ContainerStarted","Data":"7ed545424d45e920ff3e16f3b8efc23176f931a29eb6d87980057b9c4b74f912"} Oct 02 10:06:37 crc kubenswrapper[4771]: I1002 10:06:37.700329 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d8b4155-64a1-4c45-b368-72013bcdb7aa" path="/var/lib/kubelet/pods/9d8b4155-64a1-4c45-b368-72013bcdb7aa/volumes" Oct 02 10:06:39 crc kubenswrapper[4771]: E1002 10:06:39.532557 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="763b93cb9e0bd45f8673736c7da2c1d7ebc5601f38a1b419d4f03d0dea0ff226" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 02 10:06:39 crc kubenswrapper[4771]: E1002 10:06:39.534808 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="763b93cb9e0bd45f8673736c7da2c1d7ebc5601f38a1b419d4f03d0dea0ff226" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 02 10:06:39 crc kubenswrapper[4771]: E1002 10:06:39.536350 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="763b93cb9e0bd45f8673736c7da2c1d7ebc5601f38a1b419d4f03d0dea0ff226" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 02 10:06:39 crc kubenswrapper[4771]: E1002 10:06:39.536387 4771 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-868b7bfb89-mjrmr" podUID="a4bc89da-2332-412a-b7af-6056f830310a" containerName="heat-engine" Oct 02 10:06:42 crc kubenswrapper[4771]: I1002 10:06:42.508985 4771 generic.go:334] "Generic (PLEG): container finished" podID="58b8222a-a342-4b38-8aa6-47c6ea065493" containerID="a8b57ff6bd1d5ac75681e27a32751615b1abd13286e27c4275e67f3c8692579a" exitCode=0 Oct 02 10:06:42 crc kubenswrapper[4771]: I1002 10:06:42.509087 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" event={"ID":"58b8222a-a342-4b38-8aa6-47c6ea065493","Type":"ContainerDied","Data":"a8b57ff6bd1d5ac75681e27a32751615b1abd13286e27c4275e67f3c8692579a"} Oct 02 10:06:43 crc kubenswrapper[4771]: I1002 10:06:43.536090 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-l5pd4" event={"ID":"94e19f78-6425-4099-8188-8dabe4c2d9f0","Type":"ContainerStarted","Data":"ca28488797bd03e21da30abbb0b1c3cc7af3a0ee4347241a01802f98eb3a6ffc"} Oct 02 10:06:43 crc kubenswrapper[4771]: I1002 10:06:43.561909 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-l5pd4" podStartSLOduration=2.224785633 podStartE2EDuration="7.561880926s" podCreationTimestamp="2025-10-02 10:06:36 +0000 UTC" firstStartedPulling="2025-10-02 10:06:37.373877053 +0000 UTC m=+1785.021562120" lastFinishedPulling="2025-10-02 10:06:42.710972346 +0000 UTC m=+1790.358657413" observedRunningTime="2025-10-02 10:06:43.557632764 +0000 UTC m=+1791.205317841" watchObservedRunningTime="2025-10-02 10:06:43.561880926 +0000 UTC m=+1791.209566003" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.372350 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.392316 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.528154 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-ssh-key\") pod \"58b8222a-a342-4b38-8aa6-47c6ea065493\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.528340 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lj8k4\" (UniqueName: \"kubernetes.io/projected/58b8222a-a342-4b38-8aa6-47c6ea065493-kube-api-access-lj8k4\") pod \"58b8222a-a342-4b38-8aa6-47c6ea065493\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.528531 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-inventory\") pod \"58b8222a-a342-4b38-8aa6-47c6ea065493\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.528651 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-repo-setup-combined-ca-bundle\") pod \"58b8222a-a342-4b38-8aa6-47c6ea065493\" (UID: \"58b8222a-a342-4b38-8aa6-47c6ea065493\") " Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.547839 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58b8222a-a342-4b38-8aa6-47c6ea065493-kube-api-access-lj8k4" (OuterVolumeSpecName: "kube-api-access-lj8k4") pod "58b8222a-a342-4b38-8aa6-47c6ea065493" (UID: "58b8222a-a342-4b38-8aa6-47c6ea065493"). InnerVolumeSpecName "kube-api-access-lj8k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.556957 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "58b8222a-a342-4b38-8aa6-47c6ea065493" (UID: "58b8222a-a342-4b38-8aa6-47c6ea065493"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.582989 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.583242 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj" event={"ID":"58b8222a-a342-4b38-8aa6-47c6ea065493","Type":"ContainerDied","Data":"d5a13739211c2f5ff92c31dd5cc4af442483456044c2a1ee291e53f21f800e35"} Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.583271 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5a13739211c2f5ff92c31dd5cc4af442483456044c2a1ee291e53f21f800e35" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.626323 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "58b8222a-a342-4b38-8aa6-47c6ea065493" (UID: "58b8222a-a342-4b38-8aa6-47c6ea065493"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.648091 4771 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.648452 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.648481 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lj8k4\" (UniqueName: \"kubernetes.io/projected/58b8222a-a342-4b38-8aa6-47c6ea065493-kube-api-access-lj8k4\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.649850 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-inventory" (OuterVolumeSpecName: "inventory") pod "58b8222a-a342-4b38-8aa6-47c6ea065493" (UID: "58b8222a-a342-4b38-8aa6-47c6ea065493"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.736861 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.801710 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz"] Oct 02 10:06:44 crc kubenswrapper[4771]: E1002 10:06:44.803464 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58b8222a-a342-4b38-8aa6-47c6ea065493" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.803496 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="58b8222a-a342-4b38-8aa6-47c6ea065493" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.804190 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="58b8222a-a342-4b38-8aa6-47c6ea065493" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.849304 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.858540 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58b8222a-a342-4b38-8aa6-47c6ea065493-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:44 crc kubenswrapper[4771]: I1002 10:06:44.876980 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz"] Oct 02 10:06:45 crc kubenswrapper[4771]: I1002 10:06:45.066244 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7c7x\" (UniqueName: \"kubernetes.io/projected/e7004c3e-7b27-47fc-accf-7156c0fe77b0-kube-api-access-m7c7x\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbffz\" (UID: \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" Oct 02 10:06:45 crc kubenswrapper[4771]: I1002 10:06:45.066308 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7004c3e-7b27-47fc-accf-7156c0fe77b0-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbffz\" (UID: \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" Oct 02 10:06:45 crc kubenswrapper[4771]: I1002 10:06:45.066493 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7004c3e-7b27-47fc-accf-7156c0fe77b0-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbffz\" (UID: \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" Oct 02 10:06:45 crc kubenswrapper[4771]: I1002 10:06:45.170857 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7c7x\" (UniqueName: \"kubernetes.io/projected/e7004c3e-7b27-47fc-accf-7156c0fe77b0-kube-api-access-m7c7x\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbffz\" (UID: \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" Oct 02 10:06:45 crc kubenswrapper[4771]: I1002 10:06:45.170949 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7004c3e-7b27-47fc-accf-7156c0fe77b0-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbffz\" (UID: \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" Oct 02 10:06:45 crc kubenswrapper[4771]: I1002 10:06:45.171258 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7004c3e-7b27-47fc-accf-7156c0fe77b0-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbffz\" (UID: \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" Oct 02 10:06:45 crc kubenswrapper[4771]: I1002 10:06:45.176830 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7004c3e-7b27-47fc-accf-7156c0fe77b0-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbffz\" (UID: \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" Oct 02 10:06:45 crc kubenswrapper[4771]: I1002 10:06:45.186887 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7004c3e-7b27-47fc-accf-7156c0fe77b0-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbffz\" (UID: \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" Oct 02 10:06:45 crc kubenswrapper[4771]: I1002 10:06:45.195808 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7c7x\" (UniqueName: \"kubernetes.io/projected/e7004c3e-7b27-47fc-accf-7156c0fe77b0-kube-api-access-m7c7x\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbffz\" (UID: \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" Oct 02 10:06:45 crc kubenswrapper[4771]: I1002 10:06:45.489855 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" Oct 02 10:06:46 crc kubenswrapper[4771]: I1002 10:06:46.126948 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz"] Oct 02 10:06:46 crc kubenswrapper[4771]: I1002 10:06:46.606565 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" event={"ID":"e7004c3e-7b27-47fc-accf-7156c0fe77b0","Type":"ContainerStarted","Data":"54773434d5e87b6a788a3922c13890ff5d4a98da19936f83be019705775f9b2d"} Oct 02 10:06:47 crc kubenswrapper[4771]: I1002 10:06:47.619988 4771 generic.go:334] "Generic (PLEG): container finished" podID="94e19f78-6425-4099-8188-8dabe4c2d9f0" containerID="ca28488797bd03e21da30abbb0b1c3cc7af3a0ee4347241a01802f98eb3a6ffc" exitCode=0 Oct 02 10:06:47 crc kubenswrapper[4771]: I1002 10:06:47.620043 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-l5pd4" event={"ID":"94e19f78-6425-4099-8188-8dabe4c2d9f0","Type":"ContainerDied","Data":"ca28488797bd03e21da30abbb0b1c3cc7af3a0ee4347241a01802f98eb3a6ffc"} Oct 02 10:06:47 crc kubenswrapper[4771]: I1002 10:06:47.622984 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" event={"ID":"e7004c3e-7b27-47fc-accf-7156c0fe77b0","Type":"ContainerStarted","Data":"4402d2e55dd59f1394a97f22fed2d147fb0b6011b74cfcd53670400082c49d12"} Oct 02 10:06:47 crc kubenswrapper[4771]: I1002 10:06:47.668745 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" podStartSLOduration=3.197127815 podStartE2EDuration="3.668715127s" podCreationTimestamp="2025-10-02 10:06:44 +0000 UTC" firstStartedPulling="2025-10-02 10:06:46.130390884 +0000 UTC m=+1793.778075951" lastFinishedPulling="2025-10-02 10:06:46.601978176 +0000 UTC m=+1794.249663263" observedRunningTime="2025-10-02 10:06:47.657517061 +0000 UTC m=+1795.305202138" watchObservedRunningTime="2025-10-02 10:06:47.668715127 +0000 UTC m=+1795.316400194" Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.291504 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.397984 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-config-data\") pod \"94e19f78-6425-4099-8188-8dabe4c2d9f0\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.398515 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm8q7\" (UniqueName: \"kubernetes.io/projected/94e19f78-6425-4099-8188-8dabe4c2d9f0-kube-api-access-wm8q7\") pod \"94e19f78-6425-4099-8188-8dabe4c2d9f0\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.398580 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-combined-ca-bundle\") pod \"94e19f78-6425-4099-8188-8dabe4c2d9f0\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.398639 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-scripts\") pod \"94e19f78-6425-4099-8188-8dabe4c2d9f0\" (UID: \"94e19f78-6425-4099-8188-8dabe4c2d9f0\") " Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.406593 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-scripts" (OuterVolumeSpecName: "scripts") pod "94e19f78-6425-4099-8188-8dabe4c2d9f0" (UID: "94e19f78-6425-4099-8188-8dabe4c2d9f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.408074 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94e19f78-6425-4099-8188-8dabe4c2d9f0-kube-api-access-wm8q7" (OuterVolumeSpecName: "kube-api-access-wm8q7") pod "94e19f78-6425-4099-8188-8dabe4c2d9f0" (UID: "94e19f78-6425-4099-8188-8dabe4c2d9f0"). InnerVolumeSpecName "kube-api-access-wm8q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.433893 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-config-data" (OuterVolumeSpecName: "config-data") pod "94e19f78-6425-4099-8188-8dabe4c2d9f0" (UID: "94e19f78-6425-4099-8188-8dabe4c2d9f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.434976 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94e19f78-6425-4099-8188-8dabe4c2d9f0" (UID: "94e19f78-6425-4099-8188-8dabe4c2d9f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.501601 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm8q7\" (UniqueName: \"kubernetes.io/projected/94e19f78-6425-4099-8188-8dabe4c2d9f0-kube-api-access-wm8q7\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.501946 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.501960 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.501971 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94e19f78-6425-4099-8188-8dabe4c2d9f0-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:49 crc kubenswrapper[4771]: E1002 10:06:49.529107 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="763b93cb9e0bd45f8673736c7da2c1d7ebc5601f38a1b419d4f03d0dea0ff226" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 02 10:06:49 crc kubenswrapper[4771]: E1002 10:06:49.530494 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="763b93cb9e0bd45f8673736c7da2c1d7ebc5601f38a1b419d4f03d0dea0ff226" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 02 10:06:49 crc kubenswrapper[4771]: E1002 10:06:49.535722 4771 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="763b93cb9e0bd45f8673736c7da2c1d7ebc5601f38a1b419d4f03d0dea0ff226" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 02 10:06:49 crc kubenswrapper[4771]: E1002 10:06:49.535832 4771 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-868b7bfb89-mjrmr" podUID="a4bc89da-2332-412a-b7af-6056f830310a" containerName="heat-engine" Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.654818 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-l5pd4" event={"ID":"94e19f78-6425-4099-8188-8dabe4c2d9f0","Type":"ContainerDied","Data":"7ed545424d45e920ff3e16f3b8efc23176f931a29eb6d87980057b9c4b74f912"} Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.654862 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ed545424d45e920ff3e16f3b8efc23176f931a29eb6d87980057b9c4b74f912" Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.654939 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-l5pd4" Oct 02 10:06:49 crc kubenswrapper[4771]: I1002 10:06:49.682043 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:06:49 crc kubenswrapper[4771]: E1002 10:06:49.682552 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.673622 4771 generic.go:334] "Generic (PLEG): container finished" podID="a4bc89da-2332-412a-b7af-6056f830310a" containerID="763b93cb9e0bd45f8673736c7da2c1d7ebc5601f38a1b419d4f03d0dea0ff226" exitCode=0 Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.675067 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-868b7bfb89-mjrmr" event={"ID":"a4bc89da-2332-412a-b7af-6056f830310a","Type":"ContainerDied","Data":"763b93cb9e0bd45f8673736c7da2c1d7ebc5601f38a1b419d4f03d0dea0ff226"} Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.675106 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-868b7bfb89-mjrmr" event={"ID":"a4bc89da-2332-412a-b7af-6056f830310a","Type":"ContainerDied","Data":"c99213ee773c091ec2ec06c5e15dc68d76582db6e7aa50cea17319754bdfb116"} Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.675116 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c99213ee773c091ec2ec06c5e15dc68d76582db6e7aa50cea17319754bdfb116" Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.677641 4771 generic.go:334] "Generic (PLEG): container finished" podID="e7004c3e-7b27-47fc-accf-7156c0fe77b0" containerID="4402d2e55dd59f1394a97f22fed2d147fb0b6011b74cfcd53670400082c49d12" exitCode=0 Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.677765 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" event={"ID":"e7004c3e-7b27-47fc-accf-7156c0fe77b0","Type":"ContainerDied","Data":"4402d2e55dd59f1394a97f22fed2d147fb0b6011b74cfcd53670400082c49d12"} Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.749331 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.840259 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-config-data-custom\") pod \"a4bc89da-2332-412a-b7af-6056f830310a\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.840343 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-combined-ca-bundle\") pod \"a4bc89da-2332-412a-b7af-6056f830310a\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.840518 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drwsr\" (UniqueName: \"kubernetes.io/projected/a4bc89da-2332-412a-b7af-6056f830310a-kube-api-access-drwsr\") pod \"a4bc89da-2332-412a-b7af-6056f830310a\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.840586 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-config-data\") pod \"a4bc89da-2332-412a-b7af-6056f830310a\" (UID: \"a4bc89da-2332-412a-b7af-6056f830310a\") " Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.845834 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4bc89da-2332-412a-b7af-6056f830310a-kube-api-access-drwsr" (OuterVolumeSpecName: "kube-api-access-drwsr") pod "a4bc89da-2332-412a-b7af-6056f830310a" (UID: "a4bc89da-2332-412a-b7af-6056f830310a"). InnerVolumeSpecName "kube-api-access-drwsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.848634 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a4bc89da-2332-412a-b7af-6056f830310a" (UID: "a4bc89da-2332-412a-b7af-6056f830310a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.875363 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4bc89da-2332-412a-b7af-6056f830310a" (UID: "a4bc89da-2332-412a-b7af-6056f830310a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.907078 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-config-data" (OuterVolumeSpecName: "config-data") pod "a4bc89da-2332-412a-b7af-6056f830310a" (UID: "a4bc89da-2332-412a-b7af-6056f830310a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.944979 4771 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.945016 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.945027 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drwsr\" (UniqueName: \"kubernetes.io/projected/a4bc89da-2332-412a-b7af-6056f830310a-kube-api-access-drwsr\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:50 crc kubenswrapper[4771]: I1002 10:06:50.945039 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4bc89da-2332-412a-b7af-6056f830310a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:51 crc kubenswrapper[4771]: I1002 10:06:51.410676 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 02 10:06:51 crc kubenswrapper[4771]: I1002 10:06:51.411051 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-api" containerID="cri-o://581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8" gracePeriod=30 Oct 02 10:06:51 crc kubenswrapper[4771]: I1002 10:06:51.411126 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-notifier" containerID="cri-o://85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38" gracePeriod=30 Oct 02 10:06:51 crc kubenswrapper[4771]: I1002 10:06:51.411184 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-listener" containerID="cri-o://3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337" gracePeriod=30 Oct 02 10:06:51 crc kubenswrapper[4771]: I1002 10:06:51.411286 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-evaluator" containerID="cri-o://5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d" gracePeriod=30 Oct 02 10:06:51 crc kubenswrapper[4771]: I1002 10:06:51.730850 4771 generic.go:334] "Generic (PLEG): container finished" podID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerID="581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8" exitCode=0 Oct 02 10:06:51 crc kubenswrapper[4771]: I1002 10:06:51.730970 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-868b7bfb89-mjrmr" Oct 02 10:06:51 crc kubenswrapper[4771]: I1002 10:06:51.732283 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8f62371d-1395-42e3-b0e8-ec5f3439441c","Type":"ContainerDied","Data":"581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8"} Oct 02 10:06:51 crc kubenswrapper[4771]: I1002 10:06:51.784331 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-868b7bfb89-mjrmr"] Oct 02 10:06:51 crc kubenswrapper[4771]: I1002 10:06:51.795572 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-868b7bfb89-mjrmr"] Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.319431 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.408646 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7004c3e-7b27-47fc-accf-7156c0fe77b0-ssh-key\") pod \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\" (UID: \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\") " Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.408800 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7c7x\" (UniqueName: \"kubernetes.io/projected/e7004c3e-7b27-47fc-accf-7156c0fe77b0-kube-api-access-m7c7x\") pod \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\" (UID: \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\") " Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.408840 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7004c3e-7b27-47fc-accf-7156c0fe77b0-inventory\") pod \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\" (UID: \"e7004c3e-7b27-47fc-accf-7156c0fe77b0\") " Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.424027 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7004c3e-7b27-47fc-accf-7156c0fe77b0-kube-api-access-m7c7x" (OuterVolumeSpecName: "kube-api-access-m7c7x") pod "e7004c3e-7b27-47fc-accf-7156c0fe77b0" (UID: "e7004c3e-7b27-47fc-accf-7156c0fe77b0"). InnerVolumeSpecName "kube-api-access-m7c7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.505447 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7004c3e-7b27-47fc-accf-7156c0fe77b0-inventory" (OuterVolumeSpecName: "inventory") pod "e7004c3e-7b27-47fc-accf-7156c0fe77b0" (UID: "e7004c3e-7b27-47fc-accf-7156c0fe77b0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.510843 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7c7x\" (UniqueName: \"kubernetes.io/projected/e7004c3e-7b27-47fc-accf-7156c0fe77b0-kube-api-access-m7c7x\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.510871 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7004c3e-7b27-47fc-accf-7156c0fe77b0-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.519410 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7004c3e-7b27-47fc-accf-7156c0fe77b0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e7004c3e-7b27-47fc-accf-7156c0fe77b0" (UID: "e7004c3e-7b27-47fc-accf-7156c0fe77b0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.615252 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7004c3e-7b27-47fc-accf-7156c0fe77b0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.743623 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" event={"ID":"e7004c3e-7b27-47fc-accf-7156c0fe77b0","Type":"ContainerDied","Data":"54773434d5e87b6a788a3922c13890ff5d4a98da19936f83be019705775f9b2d"} Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.743680 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54773434d5e87b6a788a3922c13890ff5d4a98da19936f83be019705775f9b2d" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.743750 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbffz" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.746611 4771 generic.go:334] "Generic (PLEG): container finished" podID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerID="5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d" exitCode=0 Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.746669 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8f62371d-1395-42e3-b0e8-ec5f3439441c","Type":"ContainerDied","Data":"5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d"} Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.785818 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4"] Oct 02 10:06:52 crc kubenswrapper[4771]: E1002 10:06:52.786427 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4bc89da-2332-412a-b7af-6056f830310a" containerName="heat-engine" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.786448 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4bc89da-2332-412a-b7af-6056f830310a" containerName="heat-engine" Oct 02 10:06:52 crc kubenswrapper[4771]: E1002 10:06:52.786477 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e19f78-6425-4099-8188-8dabe4c2d9f0" containerName="aodh-db-sync" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.786491 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e19f78-6425-4099-8188-8dabe4c2d9f0" containerName="aodh-db-sync" Oct 02 10:06:52 crc kubenswrapper[4771]: E1002 10:06:52.786513 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7004c3e-7b27-47fc-accf-7156c0fe77b0" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.786523 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7004c3e-7b27-47fc-accf-7156c0fe77b0" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.786834 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="94e19f78-6425-4099-8188-8dabe4c2d9f0" containerName="aodh-db-sync" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.786856 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4bc89da-2332-412a-b7af-6056f830310a" containerName="heat-engine" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.786898 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7004c3e-7b27-47fc-accf-7156c0fe77b0" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.788661 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.791614 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.791660 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.792041 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.792453 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.805300 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4"] Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.924547 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.925330 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpplx\" (UniqueName: \"kubernetes.io/projected/4b64be94-5fae-49bf-83fd-b131c2d8aa24-kube-api-access-dpplx\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.925466 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:52 crc kubenswrapper[4771]: I1002 10:06:52.925696 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:53 crc kubenswrapper[4771]: I1002 10:06:53.028632 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpplx\" (UniqueName: \"kubernetes.io/projected/4b64be94-5fae-49bf-83fd-b131c2d8aa24-kube-api-access-dpplx\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:53 crc kubenswrapper[4771]: I1002 10:06:53.028776 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:53 crc kubenswrapper[4771]: I1002 10:06:53.028873 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:53 crc kubenswrapper[4771]: I1002 10:06:53.029165 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:53 crc kubenswrapper[4771]: I1002 10:06:53.037758 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:53 crc kubenswrapper[4771]: I1002 10:06:53.040775 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:53 crc kubenswrapper[4771]: I1002 10:06:53.042408 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:53 crc kubenswrapper[4771]: I1002 10:06:53.048300 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpplx\" (UniqueName: \"kubernetes.io/projected/4b64be94-5fae-49bf-83fd-b131c2d8aa24-kube-api-access-dpplx\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:53 crc kubenswrapper[4771]: I1002 10:06:53.113797 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:06:53 crc kubenswrapper[4771]: I1002 10:06:53.709340 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4bc89da-2332-412a-b7af-6056f830310a" path="/var/lib/kubelet/pods/a4bc89da-2332-412a-b7af-6056f830310a/volumes" Oct 02 10:06:53 crc kubenswrapper[4771]: I1002 10:06:53.712353 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4"] Oct 02 10:06:53 crc kubenswrapper[4771]: I1002 10:06:53.761115 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" event={"ID":"4b64be94-5fae-49bf-83fd-b131c2d8aa24","Type":"ContainerStarted","Data":"d9ddc8d6b7cdb6c101795882c16fc0638cd7f8412913dbe8bc48998d650f173c"} Oct 02 10:06:54 crc kubenswrapper[4771]: I1002 10:06:54.151203 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:06:54 crc kubenswrapper[4771]: I1002 10:06:54.777712 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" event={"ID":"4b64be94-5fae-49bf-83fd-b131c2d8aa24","Type":"ContainerStarted","Data":"590e70a3e05c1680abd4601a675431b65c54d2a15f196ae67464296bf7f12edc"} Oct 02 10:06:54 crc kubenswrapper[4771]: I1002 10:06:54.803268 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" podStartSLOduration=2.373075206 podStartE2EDuration="2.803245731s" podCreationTimestamp="2025-10-02 10:06:52 +0000 UTC" firstStartedPulling="2025-10-02 10:06:53.718251936 +0000 UTC m=+1801.365937003" lastFinishedPulling="2025-10-02 10:06:54.148422461 +0000 UTC m=+1801.796107528" observedRunningTime="2025-10-02 10:06:54.797449727 +0000 UTC m=+1802.445134794" watchObservedRunningTime="2025-10-02 10:06:54.803245731 +0000 UTC m=+1802.450930798" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.748568 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.794424 4771 generic.go:334] "Generic (PLEG): container finished" podID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerID="3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337" exitCode=0 Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.794470 4771 generic.go:334] "Generic (PLEG): container finished" podID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerID="85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38" exitCode=0 Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.794519 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.794520 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8f62371d-1395-42e3-b0e8-ec5f3439441c","Type":"ContainerDied","Data":"3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337"} Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.794610 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8f62371d-1395-42e3-b0e8-ec5f3439441c","Type":"ContainerDied","Data":"85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38"} Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.794623 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8f62371d-1395-42e3-b0e8-ec5f3439441c","Type":"ContainerDied","Data":"87777b2e0efe794e8660b8f467b2688298544db6a9652f3a61193c0bb4972ada"} Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.794645 4771 scope.go:117] "RemoveContainer" containerID="3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.831438 4771 scope.go:117] "RemoveContainer" containerID="85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.862040 4771 scope.go:117] "RemoveContainer" containerID="5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.890383 4771 scope.go:117] "RemoveContainer" containerID="581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.914436 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf8s6\" (UniqueName: \"kubernetes.io/projected/8f62371d-1395-42e3-b0e8-ec5f3439441c-kube-api-access-cf8s6\") pod \"8f62371d-1395-42e3-b0e8-ec5f3439441c\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.914802 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-public-tls-certs\") pod \"8f62371d-1395-42e3-b0e8-ec5f3439441c\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.920111 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-config-data\") pod \"8f62371d-1395-42e3-b0e8-ec5f3439441c\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.920430 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-combined-ca-bundle\") pod \"8f62371d-1395-42e3-b0e8-ec5f3439441c\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.920618 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-scripts\") pod \"8f62371d-1395-42e3-b0e8-ec5f3439441c\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.920872 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-internal-tls-certs\") pod \"8f62371d-1395-42e3-b0e8-ec5f3439441c\" (UID: \"8f62371d-1395-42e3-b0e8-ec5f3439441c\") " Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.926770 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-scripts" (OuterVolumeSpecName: "scripts") pod "8f62371d-1395-42e3-b0e8-ec5f3439441c" (UID: "8f62371d-1395-42e3-b0e8-ec5f3439441c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.927456 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f62371d-1395-42e3-b0e8-ec5f3439441c-kube-api-access-cf8s6" (OuterVolumeSpecName: "kube-api-access-cf8s6") pod "8f62371d-1395-42e3-b0e8-ec5f3439441c" (UID: "8f62371d-1395-42e3-b0e8-ec5f3439441c"). InnerVolumeSpecName "kube-api-access-cf8s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.931699 4771 scope.go:117] "RemoveContainer" containerID="3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337" Oct 02 10:06:55 crc kubenswrapper[4771]: E1002 10:06:55.938863 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337\": container with ID starting with 3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337 not found: ID does not exist" containerID="3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.940463 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337"} err="failed to get container status \"3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337\": rpc error: code = NotFound desc = could not find container \"3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337\": container with ID starting with 3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337 not found: ID does not exist" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.940510 4771 scope.go:117] "RemoveContainer" containerID="85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38" Oct 02 10:06:55 crc kubenswrapper[4771]: E1002 10:06:55.942494 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38\": container with ID starting with 85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38 not found: ID does not exist" containerID="85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.942555 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38"} err="failed to get container status \"85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38\": rpc error: code = NotFound desc = could not find container \"85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38\": container with ID starting with 85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38 not found: ID does not exist" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.942597 4771 scope.go:117] "RemoveContainer" containerID="5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d" Oct 02 10:06:55 crc kubenswrapper[4771]: E1002 10:06:55.961064 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d\": container with ID starting with 5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d not found: ID does not exist" containerID="5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.961112 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d"} err="failed to get container status \"5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d\": rpc error: code = NotFound desc = could not find container \"5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d\": container with ID starting with 5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d not found: ID does not exist" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.961158 4771 scope.go:117] "RemoveContainer" containerID="581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8" Oct 02 10:06:55 crc kubenswrapper[4771]: E1002 10:06:55.961514 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8\": container with ID starting with 581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8 not found: ID does not exist" containerID="581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.961588 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8"} err="failed to get container status \"581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8\": rpc error: code = NotFound desc = could not find container \"581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8\": container with ID starting with 581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8 not found: ID does not exist" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.961625 4771 scope.go:117] "RemoveContainer" containerID="3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.963080 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337"} err="failed to get container status \"3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337\": rpc error: code = NotFound desc = could not find container \"3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337\": container with ID starting with 3d919e2d067f76178e96eb3c18489d4db363d8515058e706034c2d984d508337 not found: ID does not exist" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.963110 4771 scope.go:117] "RemoveContainer" containerID="85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.963312 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38"} err="failed to get container status \"85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38\": rpc error: code = NotFound desc = could not find container \"85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38\": container with ID starting with 85c3cd6b05f095ac89585621f01164d9838d1ac706eba61620da608bc4f32b38 not found: ID does not exist" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.963332 4771 scope.go:117] "RemoveContainer" containerID="5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.963880 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d"} err="failed to get container status \"5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d\": rpc error: code = NotFound desc = could not find container \"5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d\": container with ID starting with 5bb447f5516b55b5e98fd4e6889b0517ebd3c187925b3367006b953a919b020d not found: ID does not exist" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.963902 4771 scope.go:117] "RemoveContainer" containerID="581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8" Oct 02 10:06:55 crc kubenswrapper[4771]: I1002 10:06:55.964079 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8"} err="failed to get container status \"581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8\": rpc error: code = NotFound desc = could not find container \"581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8\": container with ID starting with 581e43d5d8dee7c205cd0ddb6f0bdceec92eb5b61882fba8c226b73434ae24c8 not found: ID does not exist" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.027114 4771 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.027179 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf8s6\" (UniqueName: \"kubernetes.io/projected/8f62371d-1395-42e3-b0e8-ec5f3439441c-kube-api-access-cf8s6\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.039307 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8f62371d-1395-42e3-b0e8-ec5f3439441c" (UID: "8f62371d-1395-42e3-b0e8-ec5f3439441c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.084412 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8f62371d-1395-42e3-b0e8-ec5f3439441c" (UID: "8f62371d-1395-42e3-b0e8-ec5f3439441c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.130077 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-config-data" (OuterVolumeSpecName: "config-data") pod "8f62371d-1395-42e3-b0e8-ec5f3439441c" (UID: "8f62371d-1395-42e3-b0e8-ec5f3439441c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.133306 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f62371d-1395-42e3-b0e8-ec5f3439441c" (UID: "8f62371d-1395-42e3-b0e8-ec5f3439441c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.134344 4771 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.134385 4771 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.134398 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.134410 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f62371d-1395-42e3-b0e8-ec5f3439441c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.453648 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.469868 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.511656 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 02 10:06:56 crc kubenswrapper[4771]: E1002 10:06:56.512481 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-notifier" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.512505 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-notifier" Oct 02 10:06:56 crc kubenswrapper[4771]: E1002 10:06:56.512562 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-listener" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.512571 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-listener" Oct 02 10:06:56 crc kubenswrapper[4771]: E1002 10:06:56.512586 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-evaluator" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.512595 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-evaluator" Oct 02 10:06:56 crc kubenswrapper[4771]: E1002 10:06:56.512611 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-api" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.512618 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-api" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.512940 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-listener" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.512968 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-api" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.512981 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-evaluator" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.513001 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" containerName="aodh-notifier" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.515939 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.521423 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.521833 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.522014 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.524972 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-mnlz2" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.539354 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.543345 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.545494 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-config-data\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.545767 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-public-tls-certs\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.545928 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-internal-tls-certs\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.546179 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.546449 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4s68\" (UniqueName: \"kubernetes.io/projected/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-kube-api-access-w4s68\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.546591 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-scripts\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.658647 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-config-data\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.659299 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-public-tls-certs\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.659396 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-internal-tls-certs\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.659643 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.659818 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4s68\" (UniqueName: \"kubernetes.io/projected/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-kube-api-access-w4s68\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.659882 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-scripts\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.664324 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-scripts\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.681765 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-public-tls-certs\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.689907 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-config-data\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.701647 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.710541 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-internal-tls-certs\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.713854 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4s68\" (UniqueName: \"kubernetes.io/projected/bc2a08a4-b679-4bcc-ac8a-db9d7108b69c-kube-api-access-w4s68\") pod \"aodh-0\" (UID: \"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c\") " pod="openstack/aodh-0" Oct 02 10:06:56 crc kubenswrapper[4771]: I1002 10:06:56.841103 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 02 10:06:57 crc kubenswrapper[4771]: I1002 10:06:57.359028 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 02 10:06:57 crc kubenswrapper[4771]: I1002 10:06:57.698582 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f62371d-1395-42e3-b0e8-ec5f3439441c" path="/var/lib/kubelet/pods/8f62371d-1395-42e3-b0e8-ec5f3439441c/volumes" Oct 02 10:06:57 crc kubenswrapper[4771]: I1002 10:06:57.821550 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c","Type":"ContainerStarted","Data":"a159ba7318182eb231d9ece984aa01c7b91e41fca6ec9b38dc8fae618bde34cf"} Oct 02 10:06:58 crc kubenswrapper[4771]: I1002 10:06:58.834899 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c","Type":"ContainerStarted","Data":"f2512e71c68bab716b47cb1456e0ff21937cb8f454542323eec4e3203b9c1aa7"} Oct 02 10:06:59 crc kubenswrapper[4771]: I1002 10:06:59.851094 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c","Type":"ContainerStarted","Data":"8dfb1573e652fa75fb984963ad2134ae3f5267b390505f94422031f40225e120"} Oct 02 10:07:00 crc kubenswrapper[4771]: I1002 10:07:00.067187 4771 scope.go:117] "RemoveContainer" containerID="64a69a71b8ee1037e5918d2efd2826cac3d09f55d8d7bc0c97a2614cb0f260b3" Oct 02 10:07:00 crc kubenswrapper[4771]: I1002 10:07:00.426698 4771 scope.go:117] "RemoveContainer" containerID="206ed65503085f9fbdac82b37985708060c35a2884d9991a7324b4d264d3228f" Oct 02 10:07:00 crc kubenswrapper[4771]: I1002 10:07:00.470596 4771 scope.go:117] "RemoveContainer" containerID="e0a3421d25863823e2518b80b0336067bf482596d694069ade7893462313fb3a" Oct 02 10:07:00 crc kubenswrapper[4771]: I1002 10:07:00.495296 4771 scope.go:117] "RemoveContainer" containerID="25ace71c32266c58afbd34ba55f7b5a6a6382d12779760e38f072b19f23cb919" Oct 02 10:07:00 crc kubenswrapper[4771]: I1002 10:07:00.872020 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c","Type":"ContainerStarted","Data":"91b804854fc320819ad5378e66291bfc9622cdfb262df271c02eb65343baa013"} Oct 02 10:07:02 crc kubenswrapper[4771]: I1002 10:07:02.681605 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:07:02 crc kubenswrapper[4771]: E1002 10:07:02.682308 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:07:02 crc kubenswrapper[4771]: I1002 10:07:02.900226 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"bc2a08a4-b679-4bcc-ac8a-db9d7108b69c","Type":"ContainerStarted","Data":"4f7b2996ffe80ffc5a0a6e2cc0b297b82efcfa99af90bb2cb598979b279d0fea"} Oct 02 10:07:02 crc kubenswrapper[4771]: I1002 10:07:02.943342 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.717862335 podStartE2EDuration="6.943317687s" podCreationTimestamp="2025-10-02 10:06:56 +0000 UTC" firstStartedPulling="2025-10-02 10:06:57.373818724 +0000 UTC m=+1805.021503791" lastFinishedPulling="2025-10-02 10:07:01.599274076 +0000 UTC m=+1809.246959143" observedRunningTime="2025-10-02 10:07:02.925594498 +0000 UTC m=+1810.573279585" watchObservedRunningTime="2025-10-02 10:07:02.943317687 +0000 UTC m=+1810.591002754" Oct 02 10:07:15 crc kubenswrapper[4771]: I1002 10:07:15.681256 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:07:15 crc kubenswrapper[4771]: E1002 10:07:15.682255 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:07:28 crc kubenswrapper[4771]: I1002 10:07:28.682011 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:07:28 crc kubenswrapper[4771]: E1002 10:07:28.683225 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:07:43 crc kubenswrapper[4771]: I1002 10:07:43.708977 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:07:43 crc kubenswrapper[4771]: E1002 10:07:43.709829 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:07:56 crc kubenswrapper[4771]: I1002 10:07:56.681488 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:07:56 crc kubenswrapper[4771]: E1002 10:07:56.682536 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:08:09 crc kubenswrapper[4771]: I1002 10:08:09.683051 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:08:09 crc kubenswrapper[4771]: E1002 10:08:09.684120 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:08:22 crc kubenswrapper[4771]: I1002 10:08:22.681974 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:08:23 crc kubenswrapper[4771]: I1002 10:08:23.872814 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"ec7f5273f9ea2528fa747e0f9ebadd0f6e44b66e535ececeabf449a15093f50f"} Oct 02 10:08:32 crc kubenswrapper[4771]: I1002 10:08:32.081618 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-c52xm"] Oct 02 10:08:32 crc kubenswrapper[4771]: I1002 10:08:32.091700 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-c52xm"] Oct 02 10:08:33 crc kubenswrapper[4771]: I1002 10:08:33.040118 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-s4xgr"] Oct 02 10:08:33 crc kubenswrapper[4771]: I1002 10:08:33.054744 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-qrfml"] Oct 02 10:08:33 crc kubenswrapper[4771]: I1002 10:08:33.068314 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-s4xgr"] Oct 02 10:08:33 crc kubenswrapper[4771]: I1002 10:08:33.085457 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-qrfml"] Oct 02 10:08:33 crc kubenswrapper[4771]: I1002 10:08:33.707557 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="406c250c-ea6b-4820-b38b-6c4d5c0a16de" path="/var/lib/kubelet/pods/406c250c-ea6b-4820-b38b-6c4d5c0a16de/volumes" Oct 02 10:08:33 crc kubenswrapper[4771]: I1002 10:08:33.709260 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45eac090-969e-419f-880b-fea01daa6eae" path="/var/lib/kubelet/pods/45eac090-969e-419f-880b-fea01daa6eae/volumes" Oct 02 10:08:33 crc kubenswrapper[4771]: I1002 10:08:33.712086 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e87072ca-e829-43db-81d0-80df8e4339a7" path="/var/lib/kubelet/pods/e87072ca-e829-43db-81d0-80df8e4339a7/volumes" Oct 02 10:08:36 crc kubenswrapper[4771]: I1002 10:08:36.036196 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-xcv4v"] Oct 02 10:08:36 crc kubenswrapper[4771]: I1002 10:08:36.050288 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-xcv4v"] Oct 02 10:08:37 crc kubenswrapper[4771]: I1002 10:08:37.694490 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0b9f6fb-1605-49b6-865d-3945530f7848" path="/var/lib/kubelet/pods/b0b9f6fb-1605-49b6-865d-3945530f7848/volumes" Oct 02 10:08:44 crc kubenswrapper[4771]: I1002 10:08:44.036434 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-5bfb-account-create-sf2wc"] Oct 02 10:08:44 crc kubenswrapper[4771]: I1002 10:08:44.050972 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-5bfb-account-create-sf2wc"] Oct 02 10:08:45 crc kubenswrapper[4771]: I1002 10:08:45.699745 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8507c3c-8ba8-47e3-9529-503679e4ffa7" path="/var/lib/kubelet/pods/e8507c3c-8ba8-47e3-9529-503679e4ffa7/volumes" Oct 02 10:08:48 crc kubenswrapper[4771]: I1002 10:08:48.046474 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-fb66-account-create-q85l7"] Oct 02 10:08:48 crc kubenswrapper[4771]: I1002 10:08:48.060451 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-fb66-account-create-q85l7"] Oct 02 10:08:49 crc kubenswrapper[4771]: I1002 10:08:49.699737 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ba7c27c-2a56-4f63-9082-80935e39a0e2" path="/var/lib/kubelet/pods/1ba7c27c-2a56-4f63-9082-80935e39a0e2/volumes" Oct 02 10:08:52 crc kubenswrapper[4771]: I1002 10:08:52.031779 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-11a9-account-create-4nhgk"] Oct 02 10:08:52 crc kubenswrapper[4771]: I1002 10:08:52.044803 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-f944-account-create-752ln"] Oct 02 10:08:52 crc kubenswrapper[4771]: I1002 10:08:52.058944 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-11a9-account-create-4nhgk"] Oct 02 10:08:52 crc kubenswrapper[4771]: I1002 10:08:52.069150 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-f944-account-create-752ln"] Oct 02 10:08:53 crc kubenswrapper[4771]: I1002 10:08:53.697766 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02570cd0-0131-410c-9053-8909d7b5d95d" path="/var/lib/kubelet/pods/02570cd0-0131-410c-9053-8909d7b5d95d/volumes" Oct 02 10:08:53 crc kubenswrapper[4771]: I1002 10:08:53.698736 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97d628d6-bb40-4813-bea2-4bd87d549699" path="/var/lib/kubelet/pods/97d628d6-bb40-4813-bea2-4bd87d549699/volumes" Oct 02 10:08:56 crc kubenswrapper[4771]: I1002 10:08:56.042232 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-sd45z"] Oct 02 10:08:56 crc kubenswrapper[4771]: I1002 10:08:56.053662 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-sd45z"] Oct 02 10:08:57 crc kubenswrapper[4771]: I1002 10:08:57.696711 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8c508c8-a5cb-4b9a-866f-c695fd9f1d02" path="/var/lib/kubelet/pods/c8c508c8-a5cb-4b9a-866f-c695fd9f1d02/volumes" Oct 02 10:09:00 crc kubenswrapper[4771]: I1002 10:09:00.967626 4771 scope.go:117] "RemoveContainer" containerID="4c947da2879f48e37fb0dbc90a4af1dea5148f66c286d4eb6382e0387f8ae9de" Oct 02 10:09:01 crc kubenswrapper[4771]: I1002 10:09:01.002065 4771 scope.go:117] "RemoveContainer" containerID="d696f2e8df0ce0072ce0fb694fb19ee381869cdff489626a5ee7d213485a4e4c" Oct 02 10:09:01 crc kubenswrapper[4771]: I1002 10:09:01.068745 4771 scope.go:117] "RemoveContainer" containerID="544b74029917cc328962e02ba28141d9fca746ebba4a1d968bd3847fcef3e317" Oct 02 10:09:01 crc kubenswrapper[4771]: I1002 10:09:01.117419 4771 scope.go:117] "RemoveContainer" containerID="763b93cb9e0bd45f8673736c7da2c1d7ebc5601f38a1b419d4f03d0dea0ff226" Oct 02 10:09:01 crc kubenswrapper[4771]: I1002 10:09:01.199350 4771 scope.go:117] "RemoveContainer" containerID="7c2a21faff2c4fe7a008af7edd09446ca1f6719bbd3ae75737b5bf0415b5e8d3" Oct 02 10:09:01 crc kubenswrapper[4771]: I1002 10:09:01.224310 4771 scope.go:117] "RemoveContainer" containerID="00e777980e1c1d12746e7f51d360d96a6dee29e20ae0f960b8b35464ff777036" Oct 02 10:09:01 crc kubenswrapper[4771]: I1002 10:09:01.275967 4771 scope.go:117] "RemoveContainer" containerID="4d78210951b47065d8b907ebe3e9c60ba32b31de8da47258db41376df9122714" Oct 02 10:09:01 crc kubenswrapper[4771]: I1002 10:09:01.338433 4771 scope.go:117] "RemoveContainer" containerID="6f70a7be017545691f6e5e7c71cd0b392be013451a5736b10c5ee463ed461a1c" Oct 02 10:09:01 crc kubenswrapper[4771]: I1002 10:09:01.358314 4771 scope.go:117] "RemoveContainer" containerID="4e955b1a9de8071d2de6ea162c7d0e45ab77916019212df4ab3c9eb1b0c2d48c" Oct 02 10:09:01 crc kubenswrapper[4771]: I1002 10:09:01.395876 4771 scope.go:117] "RemoveContainer" containerID="402479870ae207145b71749e8a1a4d8700d3eeeaf425a34cc6482e569beccf0c" Oct 02 10:09:03 crc kubenswrapper[4771]: I1002 10:09:03.056540 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-f206-account-create-nlvdg"] Oct 02 10:09:03 crc kubenswrapper[4771]: I1002 10:09:03.066419 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-f206-account-create-nlvdg"] Oct 02 10:09:03 crc kubenswrapper[4771]: I1002 10:09:03.706014 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc" path="/var/lib/kubelet/pods/55933e6a-efc8-4c5d-9f6a-bf9c7e73ffdc/volumes" Oct 02 10:09:20 crc kubenswrapper[4771]: I1002 10:09:20.055333 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-5mhbt"] Oct 02 10:09:20 crc kubenswrapper[4771]: I1002 10:09:20.070645 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-krvcs"] Oct 02 10:09:20 crc kubenswrapper[4771]: I1002 10:09:20.092354 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-74swj"] Oct 02 10:09:20 crc kubenswrapper[4771]: I1002 10:09:20.105379 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-5mhbt"] Oct 02 10:09:20 crc kubenswrapper[4771]: I1002 10:09:20.148577 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-krvcs"] Oct 02 10:09:20 crc kubenswrapper[4771]: I1002 10:09:20.169615 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-74swj"] Oct 02 10:09:21 crc kubenswrapper[4771]: I1002 10:09:21.035919 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-m55ds"] Oct 02 10:09:21 crc kubenswrapper[4771]: I1002 10:09:21.046720 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-m55ds"] Oct 02 10:09:21 crc kubenswrapper[4771]: I1002 10:09:21.700481 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="099832fe-d356-4281-a374-127ce62a5fee" path="/var/lib/kubelet/pods/099832fe-d356-4281-a374-127ce62a5fee/volumes" Oct 02 10:09:21 crc kubenswrapper[4771]: I1002 10:09:21.711231 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a3923b9-976f-46e8-a672-b56698a83cbb" path="/var/lib/kubelet/pods/4a3923b9-976f-46e8-a672-b56698a83cbb/volumes" Oct 02 10:09:21 crc kubenswrapper[4771]: I1002 10:09:21.717811 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64ee0cf5-2677-4bc4-bf6d-ae15055a2689" path="/var/lib/kubelet/pods/64ee0cf5-2677-4bc4-bf6d-ae15055a2689/volumes" Oct 02 10:09:21 crc kubenswrapper[4771]: I1002 10:09:21.720885 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="709fe2b2-e133-4651-8d6a-ff41bb580e23" path="/var/lib/kubelet/pods/709fe2b2-e133-4651-8d6a-ff41bb580e23/volumes" Oct 02 10:09:52 crc kubenswrapper[4771]: I1002 10:09:52.045114 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-9938-account-create-jj42p"] Oct 02 10:09:52 crc kubenswrapper[4771]: I1002 10:09:52.057735 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-9938-account-create-jj42p"] Oct 02 10:09:53 crc kubenswrapper[4771]: I1002 10:09:53.044426 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-816a-account-create-6pk4g"] Oct 02 10:09:53 crc kubenswrapper[4771]: I1002 10:09:53.062002 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-4b06-account-create-5s4cn"] Oct 02 10:09:53 crc kubenswrapper[4771]: I1002 10:09:53.075251 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-73f2-account-create-phct2"] Oct 02 10:09:53 crc kubenswrapper[4771]: I1002 10:09:53.089270 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-816a-account-create-6pk4g"] Oct 02 10:09:53 crc kubenswrapper[4771]: I1002 10:09:53.103531 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-4b06-account-create-5s4cn"] Oct 02 10:09:53 crc kubenswrapper[4771]: I1002 10:09:53.115695 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-73f2-account-create-phct2"] Oct 02 10:09:53 crc kubenswrapper[4771]: I1002 10:09:53.698497 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cc9f109-903c-4e52-8031-56898f9128ad" path="/var/lib/kubelet/pods/6cc9f109-903c-4e52-8031-56898f9128ad/volumes" Oct 02 10:09:53 crc kubenswrapper[4771]: I1002 10:09:53.702436 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b63a764e-b703-434d-860d-7774c43035aa" path="/var/lib/kubelet/pods/b63a764e-b703-434d-860d-7774c43035aa/volumes" Oct 02 10:09:53 crc kubenswrapper[4771]: I1002 10:09:53.703584 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e215bc59-d489-4b95-8348-fe23e36f2d6e" path="/var/lib/kubelet/pods/e215bc59-d489-4b95-8348-fe23e36f2d6e/volumes" Oct 02 10:09:53 crc kubenswrapper[4771]: I1002 10:09:53.705179 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d" path="/var/lib/kubelet/pods/fd8e5b00-19a3-406b-a8fc-4a1a5cf34e6d/volumes" Oct 02 10:10:01 crc kubenswrapper[4771]: I1002 10:10:01.108934 4771 generic.go:334] "Generic (PLEG): container finished" podID="4b64be94-5fae-49bf-83fd-b131c2d8aa24" containerID="590e70a3e05c1680abd4601a675431b65c54d2a15f196ae67464296bf7f12edc" exitCode=0 Oct 02 10:10:01 crc kubenswrapper[4771]: I1002 10:10:01.109045 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" event={"ID":"4b64be94-5fae-49bf-83fd-b131c2d8aa24","Type":"ContainerDied","Data":"590e70a3e05c1680abd4601a675431b65c54d2a15f196ae67464296bf7f12edc"} Oct 02 10:10:01 crc kubenswrapper[4771]: I1002 10:10:01.736336 4771 scope.go:117] "RemoveContainer" containerID="079dd3d001178298098973e0fad451f6df3e8f546789da08a53b73571170ae9f" Oct 02 10:10:01 crc kubenswrapper[4771]: I1002 10:10:01.769090 4771 scope.go:117] "RemoveContainer" containerID="3c41f352c6503f3902610c6f2fc0bbf3c2b1eb173b36d6c2eeae980c97fea62d" Oct 02 10:10:01 crc kubenswrapper[4771]: I1002 10:10:01.831821 4771 scope.go:117] "RemoveContainer" containerID="d89bdd71c6f86666ce173a64b52637a86ceb3e85900ca707502c6fb2922f51a8" Oct 02 10:10:01 crc kubenswrapper[4771]: I1002 10:10:01.890316 4771 scope.go:117] "RemoveContainer" containerID="db924463bbf11fd168d26cff79a6e23cbd8dc5b01db6797c5cb332d55df9a30e" Oct 02 10:10:01 crc kubenswrapper[4771]: I1002 10:10:01.949956 4771 scope.go:117] "RemoveContainer" containerID="6cbac7afbdc710086e1ea377ba7123136a800ece8dad7712c4835e4829f8d2bf" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.024484 4771 scope.go:117] "RemoveContainer" containerID="d963749382478e971ca61a3a6dc2f234cad4fabff4235f72d1f4cc6c8d0890be" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.038421 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-t42dx"] Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.051041 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-t42dx"] Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.076223 4771 scope.go:117] "RemoveContainer" containerID="ec922be9092fad9973c5343b20ed6675d419ee21e23d69688a1527803a5a240b" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.105926 4771 scope.go:117] "RemoveContainer" containerID="02a52d8f511b6949ce8ca03694d385edfbd592a3aa00254acc681dca52c62ec2" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.154615 4771 scope.go:117] "RemoveContainer" containerID="b0039d8d6fa5b65a9b4acc8877a87b725d441cab8d55a36a7c6b8e7516123710" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.184353 4771 scope.go:117] "RemoveContainer" containerID="a5676c418f62a7deab378dac49f7fcd0cb147b8c3f19426fdac362056bbea66e" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.542112 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.679844 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-bootstrap-combined-ca-bundle\") pod \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.680238 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-ssh-key\") pod \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.680846 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-inventory\") pod \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.681085 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpplx\" (UniqueName: \"kubernetes.io/projected/4b64be94-5fae-49bf-83fd-b131c2d8aa24-kube-api-access-dpplx\") pod \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\" (UID: \"4b64be94-5fae-49bf-83fd-b131c2d8aa24\") " Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.686335 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "4b64be94-5fae-49bf-83fd-b131c2d8aa24" (UID: "4b64be94-5fae-49bf-83fd-b131c2d8aa24"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.687347 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b64be94-5fae-49bf-83fd-b131c2d8aa24-kube-api-access-dpplx" (OuterVolumeSpecName: "kube-api-access-dpplx") pod "4b64be94-5fae-49bf-83fd-b131c2d8aa24" (UID: "4b64be94-5fae-49bf-83fd-b131c2d8aa24"). InnerVolumeSpecName "kube-api-access-dpplx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.715351 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-inventory" (OuterVolumeSpecName: "inventory") pod "4b64be94-5fae-49bf-83fd-b131c2d8aa24" (UID: "4b64be94-5fae-49bf-83fd-b131c2d8aa24"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.716496 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4b64be94-5fae-49bf-83fd-b131c2d8aa24" (UID: "4b64be94-5fae-49bf-83fd-b131c2d8aa24"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.784881 4771 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.784925 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.784942 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b64be94-5fae-49bf-83fd-b131c2d8aa24-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:02 crc kubenswrapper[4771]: I1002 10:10:02.784955 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpplx\" (UniqueName: \"kubernetes.io/projected/4b64be94-5fae-49bf-83fd-b131c2d8aa24-kube-api-access-dpplx\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.209664 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" event={"ID":"4b64be94-5fae-49bf-83fd-b131c2d8aa24","Type":"ContainerDied","Data":"d9ddc8d6b7cdb6c101795882c16fc0638cd7f8412913dbe8bc48998d650f173c"} Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.209740 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9ddc8d6b7cdb6c101795882c16fc0638cd7f8412913dbe8bc48998d650f173c" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.209852 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.275113 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r"] Oct 02 10:10:03 crc kubenswrapper[4771]: E1002 10:10:03.275863 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b64be94-5fae-49bf-83fd-b131c2d8aa24" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.275888 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b64be94-5fae-49bf-83fd-b131c2d8aa24" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.276215 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b64be94-5fae-49bf-83fd-b131c2d8aa24" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.290787 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.307543 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r"] Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.308827 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.308896 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.309112 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.309293 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.402252 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp86t\" (UniqueName: \"kubernetes.io/projected/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-kube-api-access-zp86t\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r\" (UID: \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.402720 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r\" (UID: \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.402776 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r\" (UID: \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.504547 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r\" (UID: \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.504606 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r\" (UID: \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.504720 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp86t\" (UniqueName: \"kubernetes.io/projected/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-kube-api-access-zp86t\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r\" (UID: \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.516887 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r\" (UID: \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.517183 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r\" (UID: \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.522861 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp86t\" (UniqueName: \"kubernetes.io/projected/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-kube-api-access-zp86t\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r\" (UID: \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.644469 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" Oct 02 10:10:03 crc kubenswrapper[4771]: I1002 10:10:03.702422 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c6203ef-b851-49ba-ae59-9b62c473c497" path="/var/lib/kubelet/pods/4c6203ef-b851-49ba-ae59-9b62c473c497/volumes" Oct 02 10:10:04 crc kubenswrapper[4771]: I1002 10:10:04.171531 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r"] Oct 02 10:10:04 crc kubenswrapper[4771]: I1002 10:10:04.174987 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:10:04 crc kubenswrapper[4771]: I1002 10:10:04.234523 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" event={"ID":"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2","Type":"ContainerStarted","Data":"82563c268df07ee566dc738a1bf3f091308a6e3e108312d5a5f50aa04b7ef6ca"} Oct 02 10:10:05 crc kubenswrapper[4771]: I1002 10:10:05.247966 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" event={"ID":"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2","Type":"ContainerStarted","Data":"b60c089d3cc92819d8b93811ad9b661a358afc5be12b595749f6445458de2b14"} Oct 02 10:10:05 crc kubenswrapper[4771]: I1002 10:10:05.272968 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" podStartSLOduration=1.851265128 podStartE2EDuration="2.272951168s" podCreationTimestamp="2025-10-02 10:10:03 +0000 UTC" firstStartedPulling="2025-10-02 10:10:04.174745346 +0000 UTC m=+1991.822430403" lastFinishedPulling="2025-10-02 10:10:04.596431376 +0000 UTC m=+1992.244116443" observedRunningTime="2025-10-02 10:10:05.261579892 +0000 UTC m=+1992.909264969" watchObservedRunningTime="2025-10-02 10:10:05.272951168 +0000 UTC m=+1992.920636235" Oct 02 10:10:11 crc kubenswrapper[4771]: I1002 10:10:11.035203 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-cjqrx"] Oct 02 10:10:11 crc kubenswrapper[4771]: I1002 10:10:11.045147 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-cjqrx"] Oct 02 10:10:11 crc kubenswrapper[4771]: I1002 10:10:11.697018 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f0f7ed9-b19d-41e3-9eb3-d6431979e26a" path="/var/lib/kubelet/pods/9f0f7ed9-b19d-41e3-9eb3-d6431979e26a/volumes" Oct 02 10:10:42 crc kubenswrapper[4771]: I1002 10:10:42.146018 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:10:42 crc kubenswrapper[4771]: I1002 10:10:42.146654 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:10:57 crc kubenswrapper[4771]: I1002 10:10:57.044659 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-6gz2g"] Oct 02 10:10:57 crc kubenswrapper[4771]: I1002 10:10:57.061948 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-6gz2g"] Oct 02 10:10:57 crc kubenswrapper[4771]: I1002 10:10:57.694788 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="068a6ab1-1a57-489b-adf2-67cb4011216e" path="/var/lib/kubelet/pods/068a6ab1-1a57-489b-adf2-67cb4011216e/volumes" Oct 02 10:11:02 crc kubenswrapper[4771]: I1002 10:11:02.479559 4771 scope.go:117] "RemoveContainer" containerID="4192eb38c3d7c47dbd5f19a4a9f6cfad30fd8e5b8f2aafd5c1ac21959873cc5d" Oct 02 10:11:02 crc kubenswrapper[4771]: I1002 10:11:02.502878 4771 scope.go:117] "RemoveContainer" containerID="c689d2ca5485f23366ca643f74b034bae6592c2aa2b55dddb163cf1b16c13e69" Oct 02 10:11:02 crc kubenswrapper[4771]: I1002 10:11:02.521390 4771 scope.go:117] "RemoveContainer" containerID="fbefaa99f273d9c68cd7432a508e995273b63587f1ce7684d16c9863d6a16585" Oct 02 10:11:02 crc kubenswrapper[4771]: I1002 10:11:02.543606 4771 scope.go:117] "RemoveContainer" containerID="15ad62f3a2c0cba36a2993cfc2f36e2d4883666d4646884335e278215844e67e" Oct 02 10:11:02 crc kubenswrapper[4771]: I1002 10:11:02.612815 4771 scope.go:117] "RemoveContainer" containerID="d66e976c25ca5a30398fd5404e8f0a51209b789267b8e1d2ab18533250acb02f" Oct 02 10:11:02 crc kubenswrapper[4771]: I1002 10:11:02.642160 4771 scope.go:117] "RemoveContainer" containerID="f89a4c2ff4e398b811223f96db0542f7438cc196051779df68fe2746de06e645" Oct 02 10:11:02 crc kubenswrapper[4771]: I1002 10:11:02.706564 4771 scope.go:117] "RemoveContainer" containerID="a83520dc70d439314917715cdcf80808c5c3d3089fab074ba27a4d35e7f1053b" Oct 02 10:11:12 crc kubenswrapper[4771]: I1002 10:11:12.146147 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:11:12 crc kubenswrapper[4771]: I1002 10:11:12.146814 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:11:15 crc kubenswrapper[4771]: I1002 10:11:15.040287 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-wg2x2"] Oct 02 10:11:15 crc kubenswrapper[4771]: I1002 10:11:15.055759 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-wg2x2"] Oct 02 10:11:15 crc kubenswrapper[4771]: I1002 10:11:15.727002 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5" path="/var/lib/kubelet/pods/a7e5fca5-60e8-4b3d-b6bb-73f59b2ecab5/volumes" Oct 02 10:11:19 crc kubenswrapper[4771]: I1002 10:11:19.045727 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-5dxhb"] Oct 02 10:11:19 crc kubenswrapper[4771]: I1002 10:11:19.063449 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-5dxhb"] Oct 02 10:11:19 crc kubenswrapper[4771]: I1002 10:11:19.695944 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bc64084-015e-47d2-b20c-f241e367fe27" path="/var/lib/kubelet/pods/3bc64084-015e-47d2-b20c-f241e367fe27/volumes" Oct 02 10:11:27 crc kubenswrapper[4771]: I1002 10:11:27.032223 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-pkkwb"] Oct 02 10:11:27 crc kubenswrapper[4771]: I1002 10:11:27.045333 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-pkkwb"] Oct 02 10:11:27 crc kubenswrapper[4771]: I1002 10:11:27.695788 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a276f13-9f48-40ae-8445-ce376fce62a1" path="/var/lib/kubelet/pods/2a276f13-9f48-40ae-8445-ce376fce62a1/volumes" Oct 02 10:11:31 crc kubenswrapper[4771]: I1002 10:11:31.373829 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7s8mr"] Oct 02 10:11:31 crc kubenswrapper[4771]: I1002 10:11:31.377975 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:31 crc kubenswrapper[4771]: I1002 10:11:31.390419 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7s8mr"] Oct 02 10:11:31 crc kubenswrapper[4771]: I1002 10:11:31.519368 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f7b4423-2e95-460e-98eb-2eeb839fb923-utilities\") pod \"certified-operators-7s8mr\" (UID: \"0f7b4423-2e95-460e-98eb-2eeb839fb923\") " pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:31 crc kubenswrapper[4771]: I1002 10:11:31.519797 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f7b4423-2e95-460e-98eb-2eeb839fb923-catalog-content\") pod \"certified-operators-7s8mr\" (UID: \"0f7b4423-2e95-460e-98eb-2eeb839fb923\") " pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:31 crc kubenswrapper[4771]: I1002 10:11:31.519988 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsx5p\" (UniqueName: \"kubernetes.io/projected/0f7b4423-2e95-460e-98eb-2eeb839fb923-kube-api-access-gsx5p\") pod \"certified-operators-7s8mr\" (UID: \"0f7b4423-2e95-460e-98eb-2eeb839fb923\") " pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:31 crc kubenswrapper[4771]: I1002 10:11:31.622990 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f7b4423-2e95-460e-98eb-2eeb839fb923-catalog-content\") pod \"certified-operators-7s8mr\" (UID: \"0f7b4423-2e95-460e-98eb-2eeb839fb923\") " pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:31 crc kubenswrapper[4771]: I1002 10:11:31.623448 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsx5p\" (UniqueName: \"kubernetes.io/projected/0f7b4423-2e95-460e-98eb-2eeb839fb923-kube-api-access-gsx5p\") pod \"certified-operators-7s8mr\" (UID: \"0f7b4423-2e95-460e-98eb-2eeb839fb923\") " pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:31 crc kubenswrapper[4771]: I1002 10:11:31.623503 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f7b4423-2e95-460e-98eb-2eeb839fb923-catalog-content\") pod \"certified-operators-7s8mr\" (UID: \"0f7b4423-2e95-460e-98eb-2eeb839fb923\") " pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:31 crc kubenswrapper[4771]: I1002 10:11:31.623612 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f7b4423-2e95-460e-98eb-2eeb839fb923-utilities\") pod \"certified-operators-7s8mr\" (UID: \"0f7b4423-2e95-460e-98eb-2eeb839fb923\") " pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:31 crc kubenswrapper[4771]: I1002 10:11:31.623913 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f7b4423-2e95-460e-98eb-2eeb839fb923-utilities\") pod \"certified-operators-7s8mr\" (UID: \"0f7b4423-2e95-460e-98eb-2eeb839fb923\") " pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:31 crc kubenswrapper[4771]: I1002 10:11:31.653521 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsx5p\" (UniqueName: \"kubernetes.io/projected/0f7b4423-2e95-460e-98eb-2eeb839fb923-kube-api-access-gsx5p\") pod \"certified-operators-7s8mr\" (UID: \"0f7b4423-2e95-460e-98eb-2eeb839fb923\") " pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:31 crc kubenswrapper[4771]: I1002 10:11:31.748430 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:32 crc kubenswrapper[4771]: I1002 10:11:32.306220 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7s8mr"] Oct 02 10:11:33 crc kubenswrapper[4771]: I1002 10:11:33.204600 4771 generic.go:334] "Generic (PLEG): container finished" podID="0f7b4423-2e95-460e-98eb-2eeb839fb923" containerID="63eb0066b02b9661be82b269a658e2c920b595e121f43dd3c29fdd6fb4601381" exitCode=0 Oct 02 10:11:33 crc kubenswrapper[4771]: I1002 10:11:33.204955 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7s8mr" event={"ID":"0f7b4423-2e95-460e-98eb-2eeb839fb923","Type":"ContainerDied","Data":"63eb0066b02b9661be82b269a658e2c920b595e121f43dd3c29fdd6fb4601381"} Oct 02 10:11:33 crc kubenswrapper[4771]: I1002 10:11:33.205055 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7s8mr" event={"ID":"0f7b4423-2e95-460e-98eb-2eeb839fb923","Type":"ContainerStarted","Data":"0ffea7390de5f00ebaf9657d80c0a458d27e8abaf45aa82dfc9da726f394ccf9"} Oct 02 10:11:34 crc kubenswrapper[4771]: I1002 10:11:34.219596 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7s8mr" event={"ID":"0f7b4423-2e95-460e-98eb-2eeb839fb923","Type":"ContainerStarted","Data":"0a4259e090fb1db5cc38d46e314b03514f184af176f7df344739f4bdf6fcd60d"} Oct 02 10:11:36 crc kubenswrapper[4771]: I1002 10:11:36.241855 4771 generic.go:334] "Generic (PLEG): container finished" podID="0f7b4423-2e95-460e-98eb-2eeb839fb923" containerID="0a4259e090fb1db5cc38d46e314b03514f184af176f7df344739f4bdf6fcd60d" exitCode=0 Oct 02 10:11:36 crc kubenswrapper[4771]: I1002 10:11:36.241931 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7s8mr" event={"ID":"0f7b4423-2e95-460e-98eb-2eeb839fb923","Type":"ContainerDied","Data":"0a4259e090fb1db5cc38d46e314b03514f184af176f7df344739f4bdf6fcd60d"} Oct 02 10:11:36 crc kubenswrapper[4771]: I1002 10:11:36.759603 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qhcxc"] Oct 02 10:11:36 crc kubenswrapper[4771]: I1002 10:11:36.763323 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:11:36 crc kubenswrapper[4771]: I1002 10:11:36.772159 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qhcxc"] Oct 02 10:11:36 crc kubenswrapper[4771]: I1002 10:11:36.880964 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a195e6-8078-4c6f-ae5f-6622c7cc0795-catalog-content\") pod \"redhat-operators-qhcxc\" (UID: \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\") " pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:11:36 crc kubenswrapper[4771]: I1002 10:11:36.881086 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a195e6-8078-4c6f-ae5f-6622c7cc0795-utilities\") pod \"redhat-operators-qhcxc\" (UID: \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\") " pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:11:36 crc kubenswrapper[4771]: I1002 10:11:36.881243 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99qk4\" (UniqueName: \"kubernetes.io/projected/89a195e6-8078-4c6f-ae5f-6622c7cc0795-kube-api-access-99qk4\") pod \"redhat-operators-qhcxc\" (UID: \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\") " pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:11:36 crc kubenswrapper[4771]: I1002 10:11:36.983530 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99qk4\" (UniqueName: \"kubernetes.io/projected/89a195e6-8078-4c6f-ae5f-6622c7cc0795-kube-api-access-99qk4\") pod \"redhat-operators-qhcxc\" (UID: \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\") " pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:11:36 crc kubenswrapper[4771]: I1002 10:11:36.983691 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a195e6-8078-4c6f-ae5f-6622c7cc0795-catalog-content\") pod \"redhat-operators-qhcxc\" (UID: \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\") " pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:11:36 crc kubenswrapper[4771]: I1002 10:11:36.983820 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a195e6-8078-4c6f-ae5f-6622c7cc0795-utilities\") pod \"redhat-operators-qhcxc\" (UID: \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\") " pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:11:36 crc kubenswrapper[4771]: I1002 10:11:36.984236 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a195e6-8078-4c6f-ae5f-6622c7cc0795-catalog-content\") pod \"redhat-operators-qhcxc\" (UID: \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\") " pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:11:36 crc kubenswrapper[4771]: I1002 10:11:36.984279 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a195e6-8078-4c6f-ae5f-6622c7cc0795-utilities\") pod \"redhat-operators-qhcxc\" (UID: \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\") " pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:11:37 crc kubenswrapper[4771]: I1002 10:11:37.004626 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99qk4\" (UniqueName: \"kubernetes.io/projected/89a195e6-8078-4c6f-ae5f-6622c7cc0795-kube-api-access-99qk4\") pod \"redhat-operators-qhcxc\" (UID: \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\") " pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:11:37 crc kubenswrapper[4771]: I1002 10:11:37.139510 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:11:37 crc kubenswrapper[4771]: I1002 10:11:37.260485 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7s8mr" event={"ID":"0f7b4423-2e95-460e-98eb-2eeb839fb923","Type":"ContainerStarted","Data":"541c504c291c9ed707e4e70125b87df1bba400a486cebfffd1edddf066169926"} Oct 02 10:11:37 crc kubenswrapper[4771]: I1002 10:11:37.748282 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7s8mr" podStartSLOduration=3.276997707 podStartE2EDuration="6.748255161s" podCreationTimestamp="2025-10-02 10:11:31 +0000 UTC" firstStartedPulling="2025-10-02 10:11:33.207349559 +0000 UTC m=+2080.855034626" lastFinishedPulling="2025-10-02 10:11:36.678607012 +0000 UTC m=+2084.326292080" observedRunningTime="2025-10-02 10:11:37.286653001 +0000 UTC m=+2084.934338068" watchObservedRunningTime="2025-10-02 10:11:37.748255161 +0000 UTC m=+2085.395940228" Oct 02 10:11:37 crc kubenswrapper[4771]: I1002 10:11:37.778096 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qhcxc"] Oct 02 10:11:38 crc kubenswrapper[4771]: I1002 10:11:38.274233 4771 generic.go:334] "Generic (PLEG): container finished" podID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" containerID="e39ab515c1f517a673c0c735e829dadcb49b8556ee5ed21fdaf3c99c4c8c573e" exitCode=0 Oct 02 10:11:38 crc kubenswrapper[4771]: I1002 10:11:38.274335 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhcxc" event={"ID":"89a195e6-8078-4c6f-ae5f-6622c7cc0795","Type":"ContainerDied","Data":"e39ab515c1f517a673c0c735e829dadcb49b8556ee5ed21fdaf3c99c4c8c573e"} Oct 02 10:11:38 crc kubenswrapper[4771]: I1002 10:11:38.274611 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhcxc" event={"ID":"89a195e6-8078-4c6f-ae5f-6622c7cc0795","Type":"ContainerStarted","Data":"a9e7304a2d6481b6ea8a3122ce59d6e6f6edfdb53d1f23f54b30d289bd184c6e"} Oct 02 10:11:39 crc kubenswrapper[4771]: I1002 10:11:39.759462 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sccwr"] Oct 02 10:11:39 crc kubenswrapper[4771]: I1002 10:11:39.770688 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:39 crc kubenswrapper[4771]: I1002 10:11:39.815749 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sccwr"] Oct 02 10:11:39 crc kubenswrapper[4771]: I1002 10:11:39.959573 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-catalog-content\") pod \"community-operators-sccwr\" (UID: \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\") " pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:39 crc kubenswrapper[4771]: I1002 10:11:39.959775 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-utilities\") pod \"community-operators-sccwr\" (UID: \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\") " pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:39 crc kubenswrapper[4771]: I1002 10:11:39.959813 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj4zk\" (UniqueName: \"kubernetes.io/projected/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-kube-api-access-lj4zk\") pod \"community-operators-sccwr\" (UID: \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\") " pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:40 crc kubenswrapper[4771]: I1002 10:11:40.038757 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-f725t"] Oct 02 10:11:40 crc kubenswrapper[4771]: I1002 10:11:40.050076 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-f725t"] Oct 02 10:11:40 crc kubenswrapper[4771]: I1002 10:11:40.062150 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-utilities\") pod \"community-operators-sccwr\" (UID: \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\") " pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:40 crc kubenswrapper[4771]: I1002 10:11:40.062605 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj4zk\" (UniqueName: \"kubernetes.io/projected/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-kube-api-access-lj4zk\") pod \"community-operators-sccwr\" (UID: \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\") " pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:40 crc kubenswrapper[4771]: I1002 10:11:40.062901 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-catalog-content\") pod \"community-operators-sccwr\" (UID: \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\") " pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:40 crc kubenswrapper[4771]: I1002 10:11:40.063054 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-utilities\") pod \"community-operators-sccwr\" (UID: \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\") " pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:40 crc kubenswrapper[4771]: I1002 10:11:40.063332 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-catalog-content\") pod \"community-operators-sccwr\" (UID: \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\") " pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:40 crc kubenswrapper[4771]: I1002 10:11:40.083504 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj4zk\" (UniqueName: \"kubernetes.io/projected/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-kube-api-access-lj4zk\") pod \"community-operators-sccwr\" (UID: \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\") " pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:40 crc kubenswrapper[4771]: I1002 10:11:40.132180 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:40 crc kubenswrapper[4771]: I1002 10:11:40.315703 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhcxc" event={"ID":"89a195e6-8078-4c6f-ae5f-6622c7cc0795","Type":"ContainerStarted","Data":"ef6a4cc7c1b0170fc483fede664f80435964e75a021fd53468d2f8bd021e1347"} Oct 02 10:11:40 crc kubenswrapper[4771]: I1002 10:11:40.707624 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sccwr"] Oct 02 10:11:40 crc kubenswrapper[4771]: W1002 10:11:40.711767 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd218d28_ca60_4d4a_84ce_fc1ae2da2932.slice/crio-adab1a34da901aa9bc302184e22e455d0d3467d8b1903c22f8b4127d09a50f5b WatchSource:0}: Error finding container adab1a34da901aa9bc302184e22e455d0d3467d8b1903c22f8b4127d09a50f5b: Status 404 returned error can't find the container with id adab1a34da901aa9bc302184e22e455d0d3467d8b1903c22f8b4127d09a50f5b Oct 02 10:11:41 crc kubenswrapper[4771]: I1002 10:11:41.326813 4771 generic.go:334] "Generic (PLEG): container finished" podID="cd218d28-ca60-4d4a-84ce-fc1ae2da2932" containerID="b5707c63bc71846ebc9ca2e9e09cf00f53359a76daf2bf01ab2cb62e5420d6a4" exitCode=0 Oct 02 10:11:41 crc kubenswrapper[4771]: I1002 10:11:41.327149 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sccwr" event={"ID":"cd218d28-ca60-4d4a-84ce-fc1ae2da2932","Type":"ContainerDied","Data":"b5707c63bc71846ebc9ca2e9e09cf00f53359a76daf2bf01ab2cb62e5420d6a4"} Oct 02 10:11:41 crc kubenswrapper[4771]: I1002 10:11:41.327417 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sccwr" event={"ID":"cd218d28-ca60-4d4a-84ce-fc1ae2da2932","Type":"ContainerStarted","Data":"adab1a34da901aa9bc302184e22e455d0d3467d8b1903c22f8b4127d09a50f5b"} Oct 02 10:11:41 crc kubenswrapper[4771]: I1002 10:11:41.710014 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="975630d7-f63d-43eb-99c4-a245f4f20c24" path="/var/lib/kubelet/pods/975630d7-f63d-43eb-99c4-a245f4f20c24/volumes" Oct 02 10:11:41 crc kubenswrapper[4771]: I1002 10:11:41.748833 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:41 crc kubenswrapper[4771]: I1002 10:11:41.748904 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:42 crc kubenswrapper[4771]: I1002 10:11:42.146147 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:11:42 crc kubenswrapper[4771]: I1002 10:11:42.146471 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:11:42 crc kubenswrapper[4771]: I1002 10:11:42.146519 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 10:11:42 crc kubenswrapper[4771]: I1002 10:11:42.147338 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ec7f5273f9ea2528fa747e0f9ebadd0f6e44b66e535ececeabf449a15093f50f"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:11:42 crc kubenswrapper[4771]: I1002 10:11:42.147400 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://ec7f5273f9ea2528fa747e0f9ebadd0f6e44b66e535ececeabf449a15093f50f" gracePeriod=600 Oct 02 10:11:42 crc kubenswrapper[4771]: I1002 10:11:42.353040 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="ec7f5273f9ea2528fa747e0f9ebadd0f6e44b66e535ececeabf449a15093f50f" exitCode=0 Oct 02 10:11:42 crc kubenswrapper[4771]: I1002 10:11:42.353142 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"ec7f5273f9ea2528fa747e0f9ebadd0f6e44b66e535ececeabf449a15093f50f"} Oct 02 10:11:42 crc kubenswrapper[4771]: I1002 10:11:42.353242 4771 scope.go:117] "RemoveContainer" containerID="f4f61f738e35fdc3e228a72876fe92dfe979134cd5d102318b0a50599c3d6772" Oct 02 10:11:42 crc kubenswrapper[4771]: I1002 10:11:42.359394 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sccwr" event={"ID":"cd218d28-ca60-4d4a-84ce-fc1ae2da2932","Type":"ContainerStarted","Data":"d6e86f9a8de93e82ace4c173491fcc88c709b8a50f2230add4a7876ba493a5d9"} Oct 02 10:11:42 crc kubenswrapper[4771]: I1002 10:11:42.805949 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-7s8mr" podUID="0f7b4423-2e95-460e-98eb-2eeb839fb923" containerName="registry-server" probeResult="failure" output=< Oct 02 10:11:42 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:11:42 crc kubenswrapper[4771]: > Oct 02 10:11:43 crc kubenswrapper[4771]: I1002 10:11:43.427970 4771 generic.go:334] "Generic (PLEG): container finished" podID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" containerID="ef6a4cc7c1b0170fc483fede664f80435964e75a021fd53468d2f8bd021e1347" exitCode=0 Oct 02 10:11:43 crc kubenswrapper[4771]: I1002 10:11:43.428808 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhcxc" event={"ID":"89a195e6-8078-4c6f-ae5f-6622c7cc0795","Type":"ContainerDied","Data":"ef6a4cc7c1b0170fc483fede664f80435964e75a021fd53468d2f8bd021e1347"} Oct 02 10:11:43 crc kubenswrapper[4771]: I1002 10:11:43.439554 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2"} Oct 02 10:11:45 crc kubenswrapper[4771]: I1002 10:11:45.461594 4771 generic.go:334] "Generic (PLEG): container finished" podID="cd218d28-ca60-4d4a-84ce-fc1ae2da2932" containerID="d6e86f9a8de93e82ace4c173491fcc88c709b8a50f2230add4a7876ba493a5d9" exitCode=0 Oct 02 10:11:45 crc kubenswrapper[4771]: I1002 10:11:45.461695 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sccwr" event={"ID":"cd218d28-ca60-4d4a-84ce-fc1ae2da2932","Type":"ContainerDied","Data":"d6e86f9a8de93e82ace4c173491fcc88c709b8a50f2230add4a7876ba493a5d9"} Oct 02 10:11:45 crc kubenswrapper[4771]: I1002 10:11:45.465018 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhcxc" event={"ID":"89a195e6-8078-4c6f-ae5f-6622c7cc0795","Type":"ContainerStarted","Data":"d62c673f1fa9368fac25a4a5c3dcec90218a1e3b1940a629ad97326c37caf54e"} Oct 02 10:11:45 crc kubenswrapper[4771]: I1002 10:11:45.501349 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qhcxc" podStartSLOduration=3.43642302 podStartE2EDuration="9.501323064s" podCreationTimestamp="2025-10-02 10:11:36 +0000 UTC" firstStartedPulling="2025-10-02 10:11:38.276389136 +0000 UTC m=+2085.924074203" lastFinishedPulling="2025-10-02 10:11:44.34128918 +0000 UTC m=+2091.988974247" observedRunningTime="2025-10-02 10:11:45.498403978 +0000 UTC m=+2093.146089085" watchObservedRunningTime="2025-10-02 10:11:45.501323064 +0000 UTC m=+2093.149008131" Oct 02 10:11:46 crc kubenswrapper[4771]: I1002 10:11:46.478509 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sccwr" event={"ID":"cd218d28-ca60-4d4a-84ce-fc1ae2da2932","Type":"ContainerStarted","Data":"90c252e997ab459ae0c3e234d091577b22c4f082a8b8879c7910a5cc26aa51f1"} Oct 02 10:11:46 crc kubenswrapper[4771]: I1002 10:11:46.514296 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sccwr" podStartSLOduration=2.913643946 podStartE2EDuration="7.514275035s" podCreationTimestamp="2025-10-02 10:11:39 +0000 UTC" firstStartedPulling="2025-10-02 10:11:41.328709681 +0000 UTC m=+2088.976394748" lastFinishedPulling="2025-10-02 10:11:45.92934077 +0000 UTC m=+2093.577025837" observedRunningTime="2025-10-02 10:11:46.498640728 +0000 UTC m=+2094.146325805" watchObservedRunningTime="2025-10-02 10:11:46.514275035 +0000 UTC m=+2094.161960092" Oct 02 10:11:47 crc kubenswrapper[4771]: I1002 10:11:47.140367 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:11:47 crc kubenswrapper[4771]: I1002 10:11:47.140431 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:11:48 crc kubenswrapper[4771]: I1002 10:11:48.198720 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qhcxc" podUID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" containerName="registry-server" probeResult="failure" output=< Oct 02 10:11:48 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:11:48 crc kubenswrapper[4771]: > Oct 02 10:11:50 crc kubenswrapper[4771]: I1002 10:11:50.133056 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:50 crc kubenswrapper[4771]: I1002 10:11:50.133620 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:50 crc kubenswrapper[4771]: I1002 10:11:50.193071 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:11:51 crc kubenswrapper[4771]: I1002 10:11:51.796878 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:51 crc kubenswrapper[4771]: I1002 10:11:51.849009 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:52 crc kubenswrapper[4771]: I1002 10:11:52.031899 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7s8mr"] Oct 02 10:11:53 crc kubenswrapper[4771]: I1002 10:11:53.546822 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7s8mr" podUID="0f7b4423-2e95-460e-98eb-2eeb839fb923" containerName="registry-server" containerID="cri-o://541c504c291c9ed707e4e70125b87df1bba400a486cebfffd1edddf066169926" gracePeriod=2 Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.250161 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.344360 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f7b4423-2e95-460e-98eb-2eeb839fb923-catalog-content\") pod \"0f7b4423-2e95-460e-98eb-2eeb839fb923\" (UID: \"0f7b4423-2e95-460e-98eb-2eeb839fb923\") " Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.344600 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f7b4423-2e95-460e-98eb-2eeb839fb923-utilities\") pod \"0f7b4423-2e95-460e-98eb-2eeb839fb923\" (UID: \"0f7b4423-2e95-460e-98eb-2eeb839fb923\") " Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.344675 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsx5p\" (UniqueName: \"kubernetes.io/projected/0f7b4423-2e95-460e-98eb-2eeb839fb923-kube-api-access-gsx5p\") pod \"0f7b4423-2e95-460e-98eb-2eeb839fb923\" (UID: \"0f7b4423-2e95-460e-98eb-2eeb839fb923\") " Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.345420 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f7b4423-2e95-460e-98eb-2eeb839fb923-utilities" (OuterVolumeSpecName: "utilities") pod "0f7b4423-2e95-460e-98eb-2eeb839fb923" (UID: "0f7b4423-2e95-460e-98eb-2eeb839fb923"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.351102 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f7b4423-2e95-460e-98eb-2eeb839fb923-kube-api-access-gsx5p" (OuterVolumeSpecName: "kube-api-access-gsx5p") pod "0f7b4423-2e95-460e-98eb-2eeb839fb923" (UID: "0f7b4423-2e95-460e-98eb-2eeb839fb923"). InnerVolumeSpecName "kube-api-access-gsx5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.393418 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f7b4423-2e95-460e-98eb-2eeb839fb923-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f7b4423-2e95-460e-98eb-2eeb839fb923" (UID: "0f7b4423-2e95-460e-98eb-2eeb839fb923"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.447143 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f7b4423-2e95-460e-98eb-2eeb839fb923-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.447609 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsx5p\" (UniqueName: \"kubernetes.io/projected/0f7b4423-2e95-460e-98eb-2eeb839fb923-kube-api-access-gsx5p\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.447685 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f7b4423-2e95-460e-98eb-2eeb839fb923-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.560308 4771 generic.go:334] "Generic (PLEG): container finished" podID="0f7b4423-2e95-460e-98eb-2eeb839fb923" containerID="541c504c291c9ed707e4e70125b87df1bba400a486cebfffd1edddf066169926" exitCode=0 Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.560353 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7s8mr" event={"ID":"0f7b4423-2e95-460e-98eb-2eeb839fb923","Type":"ContainerDied","Data":"541c504c291c9ed707e4e70125b87df1bba400a486cebfffd1edddf066169926"} Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.560383 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7s8mr" event={"ID":"0f7b4423-2e95-460e-98eb-2eeb839fb923","Type":"ContainerDied","Data":"0ffea7390de5f00ebaf9657d80c0a458d27e8abaf45aa82dfc9da726f394ccf9"} Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.560401 4771 scope.go:117] "RemoveContainer" containerID="541c504c291c9ed707e4e70125b87df1bba400a486cebfffd1edddf066169926" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.561281 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7s8mr" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.602362 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7s8mr"] Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.603120 4771 scope.go:117] "RemoveContainer" containerID="0a4259e090fb1db5cc38d46e314b03514f184af176f7df344739f4bdf6fcd60d" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.611048 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7s8mr"] Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.627787 4771 scope.go:117] "RemoveContainer" containerID="63eb0066b02b9661be82b269a658e2c920b595e121f43dd3c29fdd6fb4601381" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.687681 4771 scope.go:117] "RemoveContainer" containerID="541c504c291c9ed707e4e70125b87df1bba400a486cebfffd1edddf066169926" Oct 02 10:11:54 crc kubenswrapper[4771]: E1002 10:11:54.688326 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"541c504c291c9ed707e4e70125b87df1bba400a486cebfffd1edddf066169926\": container with ID starting with 541c504c291c9ed707e4e70125b87df1bba400a486cebfffd1edddf066169926 not found: ID does not exist" containerID="541c504c291c9ed707e4e70125b87df1bba400a486cebfffd1edddf066169926" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.688367 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"541c504c291c9ed707e4e70125b87df1bba400a486cebfffd1edddf066169926"} err="failed to get container status \"541c504c291c9ed707e4e70125b87df1bba400a486cebfffd1edddf066169926\": rpc error: code = NotFound desc = could not find container \"541c504c291c9ed707e4e70125b87df1bba400a486cebfffd1edddf066169926\": container with ID starting with 541c504c291c9ed707e4e70125b87df1bba400a486cebfffd1edddf066169926 not found: ID does not exist" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.688394 4771 scope.go:117] "RemoveContainer" containerID="0a4259e090fb1db5cc38d46e314b03514f184af176f7df344739f4bdf6fcd60d" Oct 02 10:11:54 crc kubenswrapper[4771]: E1002 10:11:54.688742 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a4259e090fb1db5cc38d46e314b03514f184af176f7df344739f4bdf6fcd60d\": container with ID starting with 0a4259e090fb1db5cc38d46e314b03514f184af176f7df344739f4bdf6fcd60d not found: ID does not exist" containerID="0a4259e090fb1db5cc38d46e314b03514f184af176f7df344739f4bdf6fcd60d" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.688765 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a4259e090fb1db5cc38d46e314b03514f184af176f7df344739f4bdf6fcd60d"} err="failed to get container status \"0a4259e090fb1db5cc38d46e314b03514f184af176f7df344739f4bdf6fcd60d\": rpc error: code = NotFound desc = could not find container \"0a4259e090fb1db5cc38d46e314b03514f184af176f7df344739f4bdf6fcd60d\": container with ID starting with 0a4259e090fb1db5cc38d46e314b03514f184af176f7df344739f4bdf6fcd60d not found: ID does not exist" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.688779 4771 scope.go:117] "RemoveContainer" containerID="63eb0066b02b9661be82b269a658e2c920b595e121f43dd3c29fdd6fb4601381" Oct 02 10:11:54 crc kubenswrapper[4771]: E1002 10:11:54.689006 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63eb0066b02b9661be82b269a658e2c920b595e121f43dd3c29fdd6fb4601381\": container with ID starting with 63eb0066b02b9661be82b269a658e2c920b595e121f43dd3c29fdd6fb4601381 not found: ID does not exist" containerID="63eb0066b02b9661be82b269a658e2c920b595e121f43dd3c29fdd6fb4601381" Oct 02 10:11:54 crc kubenswrapper[4771]: I1002 10:11:54.689026 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63eb0066b02b9661be82b269a658e2c920b595e121f43dd3c29fdd6fb4601381"} err="failed to get container status \"63eb0066b02b9661be82b269a658e2c920b595e121f43dd3c29fdd6fb4601381\": rpc error: code = NotFound desc = could not find container \"63eb0066b02b9661be82b269a658e2c920b595e121f43dd3c29fdd6fb4601381\": container with ID starting with 63eb0066b02b9661be82b269a658e2c920b595e121f43dd3c29fdd6fb4601381 not found: ID does not exist" Oct 02 10:11:55 crc kubenswrapper[4771]: I1002 10:11:55.695755 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f7b4423-2e95-460e-98eb-2eeb839fb923" path="/var/lib/kubelet/pods/0f7b4423-2e95-460e-98eb-2eeb839fb923/volumes" Oct 02 10:11:58 crc kubenswrapper[4771]: I1002 10:11:58.187899 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qhcxc" podUID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" containerName="registry-server" probeResult="failure" output=< Oct 02 10:11:58 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:11:58 crc kubenswrapper[4771]: > Oct 02 10:12:00 crc kubenswrapper[4771]: I1002 10:12:00.185391 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:12:00 crc kubenswrapper[4771]: I1002 10:12:00.266596 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sccwr"] Oct 02 10:12:00 crc kubenswrapper[4771]: I1002 10:12:00.624261 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sccwr" podUID="cd218d28-ca60-4d4a-84ce-fc1ae2da2932" containerName="registry-server" containerID="cri-o://90c252e997ab459ae0c3e234d091577b22c4f082a8b8879c7910a5cc26aa51f1" gracePeriod=2 Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.198266 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.325338 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-catalog-content\") pod \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\" (UID: \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\") " Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.325672 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-utilities\") pod \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\" (UID: \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\") " Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.325795 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lj4zk\" (UniqueName: \"kubernetes.io/projected/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-kube-api-access-lj4zk\") pod \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\" (UID: \"cd218d28-ca60-4d4a-84ce-fc1ae2da2932\") " Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.326450 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-utilities" (OuterVolumeSpecName: "utilities") pod "cd218d28-ca60-4d4a-84ce-fc1ae2da2932" (UID: "cd218d28-ca60-4d4a-84ce-fc1ae2da2932"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.337041 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-kube-api-access-lj4zk" (OuterVolumeSpecName: "kube-api-access-lj4zk") pod "cd218d28-ca60-4d4a-84ce-fc1ae2da2932" (UID: "cd218d28-ca60-4d4a-84ce-fc1ae2da2932"). InnerVolumeSpecName "kube-api-access-lj4zk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.381835 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd218d28-ca60-4d4a-84ce-fc1ae2da2932" (UID: "cd218d28-ca60-4d4a-84ce-fc1ae2da2932"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.428688 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.428733 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lj4zk\" (UniqueName: \"kubernetes.io/projected/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-kube-api-access-lj4zk\") on node \"crc\" DevicePath \"\"" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.428748 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd218d28-ca60-4d4a-84ce-fc1ae2da2932-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.644083 4771 generic.go:334] "Generic (PLEG): container finished" podID="cd218d28-ca60-4d4a-84ce-fc1ae2da2932" containerID="90c252e997ab459ae0c3e234d091577b22c4f082a8b8879c7910a5cc26aa51f1" exitCode=0 Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.644527 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sccwr" event={"ID":"cd218d28-ca60-4d4a-84ce-fc1ae2da2932","Type":"ContainerDied","Data":"90c252e997ab459ae0c3e234d091577b22c4f082a8b8879c7910a5cc26aa51f1"} Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.644564 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sccwr" event={"ID":"cd218d28-ca60-4d4a-84ce-fc1ae2da2932","Type":"ContainerDied","Data":"adab1a34da901aa9bc302184e22e455d0d3467d8b1903c22f8b4127d09a50f5b"} Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.644588 4771 scope.go:117] "RemoveContainer" containerID="90c252e997ab459ae0c3e234d091577b22c4f082a8b8879c7910a5cc26aa51f1" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.644763 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sccwr" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.699647 4771 scope.go:117] "RemoveContainer" containerID="d6e86f9a8de93e82ace4c173491fcc88c709b8a50f2230add4a7876ba493a5d9" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.700012 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sccwr"] Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.705317 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sccwr"] Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.775825 4771 scope.go:117] "RemoveContainer" containerID="b5707c63bc71846ebc9ca2e9e09cf00f53359a76daf2bf01ab2cb62e5420d6a4" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.808932 4771 scope.go:117] "RemoveContainer" containerID="90c252e997ab459ae0c3e234d091577b22c4f082a8b8879c7910a5cc26aa51f1" Oct 02 10:12:01 crc kubenswrapper[4771]: E1002 10:12:01.809563 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90c252e997ab459ae0c3e234d091577b22c4f082a8b8879c7910a5cc26aa51f1\": container with ID starting with 90c252e997ab459ae0c3e234d091577b22c4f082a8b8879c7910a5cc26aa51f1 not found: ID does not exist" containerID="90c252e997ab459ae0c3e234d091577b22c4f082a8b8879c7910a5cc26aa51f1" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.809601 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90c252e997ab459ae0c3e234d091577b22c4f082a8b8879c7910a5cc26aa51f1"} err="failed to get container status \"90c252e997ab459ae0c3e234d091577b22c4f082a8b8879c7910a5cc26aa51f1\": rpc error: code = NotFound desc = could not find container \"90c252e997ab459ae0c3e234d091577b22c4f082a8b8879c7910a5cc26aa51f1\": container with ID starting with 90c252e997ab459ae0c3e234d091577b22c4f082a8b8879c7910a5cc26aa51f1 not found: ID does not exist" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.809633 4771 scope.go:117] "RemoveContainer" containerID="d6e86f9a8de93e82ace4c173491fcc88c709b8a50f2230add4a7876ba493a5d9" Oct 02 10:12:01 crc kubenswrapper[4771]: E1002 10:12:01.810105 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6e86f9a8de93e82ace4c173491fcc88c709b8a50f2230add4a7876ba493a5d9\": container with ID starting with d6e86f9a8de93e82ace4c173491fcc88c709b8a50f2230add4a7876ba493a5d9 not found: ID does not exist" containerID="d6e86f9a8de93e82ace4c173491fcc88c709b8a50f2230add4a7876ba493a5d9" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.810190 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6e86f9a8de93e82ace4c173491fcc88c709b8a50f2230add4a7876ba493a5d9"} err="failed to get container status \"d6e86f9a8de93e82ace4c173491fcc88c709b8a50f2230add4a7876ba493a5d9\": rpc error: code = NotFound desc = could not find container \"d6e86f9a8de93e82ace4c173491fcc88c709b8a50f2230add4a7876ba493a5d9\": container with ID starting with d6e86f9a8de93e82ace4c173491fcc88c709b8a50f2230add4a7876ba493a5d9 not found: ID does not exist" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.810242 4771 scope.go:117] "RemoveContainer" containerID="b5707c63bc71846ebc9ca2e9e09cf00f53359a76daf2bf01ab2cb62e5420d6a4" Oct 02 10:12:01 crc kubenswrapper[4771]: E1002 10:12:01.810595 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5707c63bc71846ebc9ca2e9e09cf00f53359a76daf2bf01ab2cb62e5420d6a4\": container with ID starting with b5707c63bc71846ebc9ca2e9e09cf00f53359a76daf2bf01ab2cb62e5420d6a4 not found: ID does not exist" containerID="b5707c63bc71846ebc9ca2e9e09cf00f53359a76daf2bf01ab2cb62e5420d6a4" Oct 02 10:12:01 crc kubenswrapper[4771]: I1002 10:12:01.810625 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5707c63bc71846ebc9ca2e9e09cf00f53359a76daf2bf01ab2cb62e5420d6a4"} err="failed to get container status \"b5707c63bc71846ebc9ca2e9e09cf00f53359a76daf2bf01ab2cb62e5420d6a4\": rpc error: code = NotFound desc = could not find container \"b5707c63bc71846ebc9ca2e9e09cf00f53359a76daf2bf01ab2cb62e5420d6a4\": container with ID starting with b5707c63bc71846ebc9ca2e9e09cf00f53359a76daf2bf01ab2cb62e5420d6a4 not found: ID does not exist" Oct 02 10:12:02 crc kubenswrapper[4771]: I1002 10:12:02.862808 4771 scope.go:117] "RemoveContainer" containerID="0acd8c30b93bd4b48f322f37ba31cfee0224365b213479f44020970df277122b" Oct 02 10:12:02 crc kubenswrapper[4771]: I1002 10:12:02.907506 4771 scope.go:117] "RemoveContainer" containerID="f001212e1a5ae1a73fad0563d8f9c34c580f6d46c6321e4e686f1062692a8922" Oct 02 10:12:02 crc kubenswrapper[4771]: I1002 10:12:02.967946 4771 scope.go:117] "RemoveContainer" containerID="8badc56bf937b29d49b7199b137adf1a05ea52645f9884aff25076cc4a12b257" Oct 02 10:12:03 crc kubenswrapper[4771]: I1002 10:12:03.017584 4771 scope.go:117] "RemoveContainer" containerID="0bf3f374add61e91e09dea2502ed16444da7aec47589d12eb1132079e99b1dec" Oct 02 10:12:03 crc kubenswrapper[4771]: I1002 10:12:03.699583 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd218d28-ca60-4d4a-84ce-fc1ae2da2932" path="/var/lib/kubelet/pods/cd218d28-ca60-4d4a-84ce-fc1ae2da2932/volumes" Oct 02 10:12:07 crc kubenswrapper[4771]: I1002 10:12:07.193041 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:12:07 crc kubenswrapper[4771]: I1002 10:12:07.251447 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:12:07 crc kubenswrapper[4771]: I1002 10:12:07.956939 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qhcxc"] Oct 02 10:12:08 crc kubenswrapper[4771]: I1002 10:12:08.722725 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qhcxc" podUID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" containerName="registry-server" containerID="cri-o://d62c673f1fa9368fac25a4a5c3dcec90218a1e3b1940a629ad97326c37caf54e" gracePeriod=2 Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.395412 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.537481 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a195e6-8078-4c6f-ae5f-6622c7cc0795-utilities\") pod \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\" (UID: \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\") " Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.538099 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99qk4\" (UniqueName: \"kubernetes.io/projected/89a195e6-8078-4c6f-ae5f-6622c7cc0795-kube-api-access-99qk4\") pod \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\" (UID: \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\") " Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.538354 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a195e6-8078-4c6f-ae5f-6622c7cc0795-catalog-content\") pod \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\" (UID: \"89a195e6-8078-4c6f-ae5f-6622c7cc0795\") " Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.538494 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89a195e6-8078-4c6f-ae5f-6622c7cc0795-utilities" (OuterVolumeSpecName: "utilities") pod "89a195e6-8078-4c6f-ae5f-6622c7cc0795" (UID: "89a195e6-8078-4c6f-ae5f-6622c7cc0795"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.539113 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a195e6-8078-4c6f-ae5f-6622c7cc0795-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.546345 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89a195e6-8078-4c6f-ae5f-6622c7cc0795-kube-api-access-99qk4" (OuterVolumeSpecName: "kube-api-access-99qk4") pod "89a195e6-8078-4c6f-ae5f-6622c7cc0795" (UID: "89a195e6-8078-4c6f-ae5f-6622c7cc0795"). InnerVolumeSpecName "kube-api-access-99qk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.634371 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89a195e6-8078-4c6f-ae5f-6622c7cc0795-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89a195e6-8078-4c6f-ae5f-6622c7cc0795" (UID: "89a195e6-8078-4c6f-ae5f-6622c7cc0795"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.641718 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99qk4\" (UniqueName: \"kubernetes.io/projected/89a195e6-8078-4c6f-ae5f-6622c7cc0795-kube-api-access-99qk4\") on node \"crc\" DevicePath \"\"" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.641779 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a195e6-8078-4c6f-ae5f-6622c7cc0795-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.737168 4771 generic.go:334] "Generic (PLEG): container finished" podID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" containerID="d62c673f1fa9368fac25a4a5c3dcec90218a1e3b1940a629ad97326c37caf54e" exitCode=0 Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.737229 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhcxc" event={"ID":"89a195e6-8078-4c6f-ae5f-6622c7cc0795","Type":"ContainerDied","Data":"d62c673f1fa9368fac25a4a5c3dcec90218a1e3b1940a629ad97326c37caf54e"} Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.738092 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhcxc" event={"ID":"89a195e6-8078-4c6f-ae5f-6622c7cc0795","Type":"ContainerDied","Data":"a9e7304a2d6481b6ea8a3122ce59d6e6f6edfdb53d1f23f54b30d289bd184c6e"} Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.737265 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhcxc" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.738160 4771 scope.go:117] "RemoveContainer" containerID="d62c673f1fa9368fac25a4a5c3dcec90218a1e3b1940a629ad97326c37caf54e" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.769107 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qhcxc"] Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.770675 4771 scope.go:117] "RemoveContainer" containerID="ef6a4cc7c1b0170fc483fede664f80435964e75a021fd53468d2f8bd021e1347" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.780146 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qhcxc"] Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.804847 4771 scope.go:117] "RemoveContainer" containerID="e39ab515c1f517a673c0c735e829dadcb49b8556ee5ed21fdaf3c99c4c8c573e" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.884168 4771 scope.go:117] "RemoveContainer" containerID="d62c673f1fa9368fac25a4a5c3dcec90218a1e3b1940a629ad97326c37caf54e" Oct 02 10:12:09 crc kubenswrapper[4771]: E1002 10:12:09.885083 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d62c673f1fa9368fac25a4a5c3dcec90218a1e3b1940a629ad97326c37caf54e\": container with ID starting with d62c673f1fa9368fac25a4a5c3dcec90218a1e3b1940a629ad97326c37caf54e not found: ID does not exist" containerID="d62c673f1fa9368fac25a4a5c3dcec90218a1e3b1940a629ad97326c37caf54e" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.885153 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62c673f1fa9368fac25a4a5c3dcec90218a1e3b1940a629ad97326c37caf54e"} err="failed to get container status \"d62c673f1fa9368fac25a4a5c3dcec90218a1e3b1940a629ad97326c37caf54e\": rpc error: code = NotFound desc = could not find container \"d62c673f1fa9368fac25a4a5c3dcec90218a1e3b1940a629ad97326c37caf54e\": container with ID starting with d62c673f1fa9368fac25a4a5c3dcec90218a1e3b1940a629ad97326c37caf54e not found: ID does not exist" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.885187 4771 scope.go:117] "RemoveContainer" containerID="ef6a4cc7c1b0170fc483fede664f80435964e75a021fd53468d2f8bd021e1347" Oct 02 10:12:09 crc kubenswrapper[4771]: E1002 10:12:09.885552 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef6a4cc7c1b0170fc483fede664f80435964e75a021fd53468d2f8bd021e1347\": container with ID starting with ef6a4cc7c1b0170fc483fede664f80435964e75a021fd53468d2f8bd021e1347 not found: ID does not exist" containerID="ef6a4cc7c1b0170fc483fede664f80435964e75a021fd53468d2f8bd021e1347" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.885599 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef6a4cc7c1b0170fc483fede664f80435964e75a021fd53468d2f8bd021e1347"} err="failed to get container status \"ef6a4cc7c1b0170fc483fede664f80435964e75a021fd53468d2f8bd021e1347\": rpc error: code = NotFound desc = could not find container \"ef6a4cc7c1b0170fc483fede664f80435964e75a021fd53468d2f8bd021e1347\": container with ID starting with ef6a4cc7c1b0170fc483fede664f80435964e75a021fd53468d2f8bd021e1347 not found: ID does not exist" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.885617 4771 scope.go:117] "RemoveContainer" containerID="e39ab515c1f517a673c0c735e829dadcb49b8556ee5ed21fdaf3c99c4c8c573e" Oct 02 10:12:09 crc kubenswrapper[4771]: E1002 10:12:09.885983 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e39ab515c1f517a673c0c735e829dadcb49b8556ee5ed21fdaf3c99c4c8c573e\": container with ID starting with e39ab515c1f517a673c0c735e829dadcb49b8556ee5ed21fdaf3c99c4c8c573e not found: ID does not exist" containerID="e39ab515c1f517a673c0c735e829dadcb49b8556ee5ed21fdaf3c99c4c8c573e" Oct 02 10:12:09 crc kubenswrapper[4771]: I1002 10:12:09.886031 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e39ab515c1f517a673c0c735e829dadcb49b8556ee5ed21fdaf3c99c4c8c573e"} err="failed to get container status \"e39ab515c1f517a673c0c735e829dadcb49b8556ee5ed21fdaf3c99c4c8c573e\": rpc error: code = NotFound desc = could not find container \"e39ab515c1f517a673c0c735e829dadcb49b8556ee5ed21fdaf3c99c4c8c573e\": container with ID starting with e39ab515c1f517a673c0c735e829dadcb49b8556ee5ed21fdaf3c99c4c8c573e not found: ID does not exist" Oct 02 10:12:11 crc kubenswrapper[4771]: I1002 10:12:11.698181 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" path="/var/lib/kubelet/pods/89a195e6-8078-4c6f-ae5f-6622c7cc0795/volumes" Oct 02 10:12:17 crc kubenswrapper[4771]: I1002 10:12:17.821150 4771 generic.go:334] "Generic (PLEG): container finished" podID="7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2" containerID="b60c089d3cc92819d8b93811ad9b661a358afc5be12b595749f6445458de2b14" exitCode=0 Oct 02 10:12:17 crc kubenswrapper[4771]: I1002 10:12:17.821270 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" event={"ID":"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2","Type":"ContainerDied","Data":"b60c089d3cc92819d8b93811ad9b661a358afc5be12b595749f6445458de2b14"} Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.357452 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.481476 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-ssh-key\") pod \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\" (UID: \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\") " Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.481567 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-inventory\") pod \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\" (UID: \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\") " Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.481755 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp86t\" (UniqueName: \"kubernetes.io/projected/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-kube-api-access-zp86t\") pod \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\" (UID: \"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2\") " Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.491698 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-kube-api-access-zp86t" (OuterVolumeSpecName: "kube-api-access-zp86t") pod "7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2" (UID: "7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2"). InnerVolumeSpecName "kube-api-access-zp86t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.517276 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-inventory" (OuterVolumeSpecName: "inventory") pod "7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2" (UID: "7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.524171 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2" (UID: "7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.584434 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.584480 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp86t\" (UniqueName: \"kubernetes.io/projected/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-kube-api-access-zp86t\") on node \"crc\" DevicePath \"\"" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.584495 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.843380 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" event={"ID":"7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2","Type":"ContainerDied","Data":"82563c268df07ee566dc738a1bf3f091308a6e3e108312d5a5f50aa04b7ef6ca"} Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.843735 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82563c268df07ee566dc738a1bf3f091308a6e3e108312d5a5f50aa04b7ef6ca" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.843477 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.938218 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8"] Oct 02 10:12:19 crc kubenswrapper[4771]: E1002 10:12:19.938866 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" containerName="registry-server" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.938893 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" containerName="registry-server" Oct 02 10:12:19 crc kubenswrapper[4771]: E1002 10:12:19.938913 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f7b4423-2e95-460e-98eb-2eeb839fb923" containerName="extract-utilities" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.938921 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f7b4423-2e95-460e-98eb-2eeb839fb923" containerName="extract-utilities" Oct 02 10:12:19 crc kubenswrapper[4771]: E1002 10:12:19.938958 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" containerName="extract-content" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.938968 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" containerName="extract-content" Oct 02 10:12:19 crc kubenswrapper[4771]: E1002 10:12:19.938986 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd218d28-ca60-4d4a-84ce-fc1ae2da2932" containerName="extract-utilities" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.939013 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd218d28-ca60-4d4a-84ce-fc1ae2da2932" containerName="extract-utilities" Oct 02 10:12:19 crc kubenswrapper[4771]: E1002 10:12:19.939026 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" containerName="extract-utilities" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.939037 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" containerName="extract-utilities" Oct 02 10:12:19 crc kubenswrapper[4771]: E1002 10:12:19.939348 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.939360 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 10:12:19 crc kubenswrapper[4771]: E1002 10:12:19.939373 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd218d28-ca60-4d4a-84ce-fc1ae2da2932" containerName="registry-server" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.939380 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd218d28-ca60-4d4a-84ce-fc1ae2da2932" containerName="registry-server" Oct 02 10:12:19 crc kubenswrapper[4771]: E1002 10:12:19.939397 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f7b4423-2e95-460e-98eb-2eeb839fb923" containerName="registry-server" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.939404 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f7b4423-2e95-460e-98eb-2eeb839fb923" containerName="registry-server" Oct 02 10:12:19 crc kubenswrapper[4771]: E1002 10:12:19.939426 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f7b4423-2e95-460e-98eb-2eeb839fb923" containerName="extract-content" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.939434 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f7b4423-2e95-460e-98eb-2eeb839fb923" containerName="extract-content" Oct 02 10:12:19 crc kubenswrapper[4771]: E1002 10:12:19.939457 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd218d28-ca60-4d4a-84ce-fc1ae2da2932" containerName="extract-content" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.939465 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd218d28-ca60-4d4a-84ce-fc1ae2da2932" containerName="extract-content" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.939749 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.939779 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd218d28-ca60-4d4a-84ce-fc1ae2da2932" containerName="registry-server" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.939800 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="89a195e6-8078-4c6f-ae5f-6622c7cc0795" containerName="registry-server" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.939824 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f7b4423-2e95-460e-98eb-2eeb839fb923" containerName="registry-server" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.940913 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.945411 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.949330 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.949359 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.949342 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:12:19 crc kubenswrapper[4771]: I1002 10:12:19.955978 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8"] Oct 02 10:12:20 crc kubenswrapper[4771]: I1002 10:12:20.094709 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d368299-6a4f-4493-8460-3fc0e53f861c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8\" (UID: \"9d368299-6a4f-4493-8460-3fc0e53f861c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" Oct 02 10:12:20 crc kubenswrapper[4771]: I1002 10:12:20.094791 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2qdm\" (UniqueName: \"kubernetes.io/projected/9d368299-6a4f-4493-8460-3fc0e53f861c-kube-api-access-z2qdm\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8\" (UID: \"9d368299-6a4f-4493-8460-3fc0e53f861c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" Oct 02 10:12:20 crc kubenswrapper[4771]: I1002 10:12:20.094866 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d368299-6a4f-4493-8460-3fc0e53f861c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8\" (UID: \"9d368299-6a4f-4493-8460-3fc0e53f861c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" Oct 02 10:12:20 crc kubenswrapper[4771]: I1002 10:12:20.197645 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d368299-6a4f-4493-8460-3fc0e53f861c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8\" (UID: \"9d368299-6a4f-4493-8460-3fc0e53f861c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" Oct 02 10:12:20 crc kubenswrapper[4771]: I1002 10:12:20.197819 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2qdm\" (UniqueName: \"kubernetes.io/projected/9d368299-6a4f-4493-8460-3fc0e53f861c-kube-api-access-z2qdm\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8\" (UID: \"9d368299-6a4f-4493-8460-3fc0e53f861c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" Oct 02 10:12:20 crc kubenswrapper[4771]: I1002 10:12:20.197915 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d368299-6a4f-4493-8460-3fc0e53f861c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8\" (UID: \"9d368299-6a4f-4493-8460-3fc0e53f861c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" Oct 02 10:12:20 crc kubenswrapper[4771]: I1002 10:12:20.202856 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d368299-6a4f-4493-8460-3fc0e53f861c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8\" (UID: \"9d368299-6a4f-4493-8460-3fc0e53f861c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" Oct 02 10:12:20 crc kubenswrapper[4771]: I1002 10:12:20.214076 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d368299-6a4f-4493-8460-3fc0e53f861c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8\" (UID: \"9d368299-6a4f-4493-8460-3fc0e53f861c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" Oct 02 10:12:20 crc kubenswrapper[4771]: I1002 10:12:20.215480 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2qdm\" (UniqueName: \"kubernetes.io/projected/9d368299-6a4f-4493-8460-3fc0e53f861c-kube-api-access-z2qdm\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8\" (UID: \"9d368299-6a4f-4493-8460-3fc0e53f861c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" Oct 02 10:12:20 crc kubenswrapper[4771]: I1002 10:12:20.266218 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" Oct 02 10:12:20 crc kubenswrapper[4771]: I1002 10:12:20.841388 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8"] Oct 02 10:12:20 crc kubenswrapper[4771]: I1002 10:12:20.884338 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" event={"ID":"9d368299-6a4f-4493-8460-3fc0e53f861c","Type":"ContainerStarted","Data":"17dbe2fdea10ec4c01edc5a813fc14872ee845bef88fee276bbb8e391ddf8771"} Oct 02 10:12:22 crc kubenswrapper[4771]: I1002 10:12:22.908947 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" event={"ID":"9d368299-6a4f-4493-8460-3fc0e53f861c","Type":"ContainerStarted","Data":"25168f71bbc5ad61edcd168b93a6b7fb36950066ddb0daec9040c61e2d2f15b5"} Oct 02 10:12:22 crc kubenswrapper[4771]: I1002 10:12:22.931866 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" podStartSLOduration=2.766674578 podStartE2EDuration="3.931846397s" podCreationTimestamp="2025-10-02 10:12:19 +0000 UTC" firstStartedPulling="2025-10-02 10:12:20.848348823 +0000 UTC m=+2128.496033880" lastFinishedPulling="2025-10-02 10:12:22.013520632 +0000 UTC m=+2129.661205699" observedRunningTime="2025-10-02 10:12:22.926817766 +0000 UTC m=+2130.574502833" watchObservedRunningTime="2025-10-02 10:12:22.931846397 +0000 UTC m=+2130.579531464" Oct 02 10:12:27 crc kubenswrapper[4771]: I1002 10:12:27.042521 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-lpdl4"] Oct 02 10:12:27 crc kubenswrapper[4771]: I1002 10:12:27.053844 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-qrkm9"] Oct 02 10:12:27 crc kubenswrapper[4771]: I1002 10:12:27.064892 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-fhkn7"] Oct 02 10:12:27 crc kubenswrapper[4771]: I1002 10:12:27.075341 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-qrkm9"] Oct 02 10:12:27 crc kubenswrapper[4771]: I1002 10:12:27.084767 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-lpdl4"] Oct 02 10:12:27 crc kubenswrapper[4771]: I1002 10:12:27.094697 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-fhkn7"] Oct 02 10:12:27 crc kubenswrapper[4771]: I1002 10:12:27.694597 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cb5981a-5bc8-4e20-b6df-77d1e5e02674" path="/var/lib/kubelet/pods/5cb5981a-5bc8-4e20-b6df-77d1e5e02674/volumes" Oct 02 10:12:27 crc kubenswrapper[4771]: I1002 10:12:27.695511 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6186a50e-d587-4972-97bf-4bb6a7a13c83" path="/var/lib/kubelet/pods/6186a50e-d587-4972-97bf-4bb6a7a13c83/volumes" Oct 02 10:12:27 crc kubenswrapper[4771]: I1002 10:12:27.696071 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6e71c3b-c209-4585-a6d4-e951dbf21b42" path="/var/lib/kubelet/pods/a6e71c3b-c209-4585-a6d4-e951dbf21b42/volumes" Oct 02 10:12:36 crc kubenswrapper[4771]: I1002 10:12:36.036563 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-fc64-account-create-c4fs5"] Oct 02 10:12:36 crc kubenswrapper[4771]: I1002 10:12:36.051277 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-053a-account-create-jt258"] Oct 02 10:12:36 crc kubenswrapper[4771]: I1002 10:12:36.061856 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0ebb-account-create-sv52h"] Oct 02 10:12:36 crc kubenswrapper[4771]: I1002 10:12:36.070209 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-fc64-account-create-c4fs5"] Oct 02 10:12:36 crc kubenswrapper[4771]: I1002 10:12:36.080454 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-053a-account-create-jt258"] Oct 02 10:12:36 crc kubenswrapper[4771]: I1002 10:12:36.089967 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0ebb-account-create-sv52h"] Oct 02 10:12:37 crc kubenswrapper[4771]: I1002 10:12:37.696218 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cb58c29-0f39-4260-85ce-9a93e3b90d99" path="/var/lib/kubelet/pods/2cb58c29-0f39-4260-85ce-9a93e3b90d99/volumes" Oct 02 10:12:37 crc kubenswrapper[4771]: I1002 10:12:37.697145 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bbca829-8863-481a-8ff3-661dc1bc9726" path="/var/lib/kubelet/pods/4bbca829-8863-481a-8ff3-661dc1bc9726/volumes" Oct 02 10:12:37 crc kubenswrapper[4771]: I1002 10:12:37.697803 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d3458e9-340f-43bf-a070-7e3ec1250176" path="/var/lib/kubelet/pods/8d3458e9-340f-43bf-a070-7e3ec1250176/volumes" Oct 02 10:13:03 crc kubenswrapper[4771]: I1002 10:13:03.210557 4771 scope.go:117] "RemoveContainer" containerID="453ad515e7c0ffc2b998189de46236c802c1791795cd8dd14d6bf8f2b7e3a0ee" Oct 02 10:13:03 crc kubenswrapper[4771]: I1002 10:13:03.238053 4771 scope.go:117] "RemoveContainer" containerID="de192baf53497d82961fd1234b70f275cb41f93639d15ee54a1f439dead69522" Oct 02 10:13:03 crc kubenswrapper[4771]: I1002 10:13:03.304959 4771 scope.go:117] "RemoveContainer" containerID="ddf68268fe2909555075263c92dc36a09e66dd98d090846f1f252ee4442c17d1" Oct 02 10:13:03 crc kubenswrapper[4771]: I1002 10:13:03.355387 4771 scope.go:117] "RemoveContainer" containerID="2830b607f8ef44d30d755763b6bd0ebe201f45c32fd97f40a505528eea8a69e2" Oct 02 10:13:03 crc kubenswrapper[4771]: I1002 10:13:03.411175 4771 scope.go:117] "RemoveContainer" containerID="f8a1dcc67e400ba67f2ef3eb32afd6f507f70cefc0023657d10303fcc00ecf81" Oct 02 10:13:03 crc kubenswrapper[4771]: I1002 10:13:03.462537 4771 scope.go:117] "RemoveContainer" containerID="251074125b8b3d0d61da59671183fc3b07b65f2ffedca9cf5ee0a3d134d74a7c" Oct 02 10:13:16 crc kubenswrapper[4771]: I1002 10:13:16.056568 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6f8xc"] Oct 02 10:13:16 crc kubenswrapper[4771]: I1002 10:13:16.070826 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6f8xc"] Oct 02 10:13:17 crc kubenswrapper[4771]: I1002 10:13:17.696700 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d32f933a-a888-44bf-8028-19623221e336" path="/var/lib/kubelet/pods/d32f933a-a888-44bf-8028-19623221e336/volumes" Oct 02 10:13:23 crc kubenswrapper[4771]: I1002 10:13:23.034674 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-bgtxq"] Oct 02 10:13:23 crc kubenswrapper[4771]: I1002 10:13:23.044319 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-bgtxq"] Oct 02 10:13:23 crc kubenswrapper[4771]: I1002 10:13:23.704733 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d02770b-766e-488b-a144-dd9d8ecc0653" path="/var/lib/kubelet/pods/8d02770b-766e-488b-a144-dd9d8ecc0653/volumes" Oct 02 10:13:35 crc kubenswrapper[4771]: I1002 10:13:35.031532 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-afdd-account-create-5znqf"] Oct 02 10:13:35 crc kubenswrapper[4771]: I1002 10:13:35.044069 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-afdd-account-create-5znqf"] Oct 02 10:13:35 crc kubenswrapper[4771]: I1002 10:13:35.698418 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01978d22-b875-4359-8447-e96168822a2a" path="/var/lib/kubelet/pods/01978d22-b875-4359-8447-e96168822a2a/volumes" Oct 02 10:13:40 crc kubenswrapper[4771]: I1002 10:13:40.779643 4771 generic.go:334] "Generic (PLEG): container finished" podID="9d368299-6a4f-4493-8460-3fc0e53f861c" containerID="25168f71bbc5ad61edcd168b93a6b7fb36950066ddb0daec9040c61e2d2f15b5" exitCode=0 Oct 02 10:13:40 crc kubenswrapper[4771]: I1002 10:13:40.779709 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" event={"ID":"9d368299-6a4f-4493-8460-3fc0e53f861c","Type":"ContainerDied","Data":"25168f71bbc5ad61edcd168b93a6b7fb36950066ddb0daec9040c61e2d2f15b5"} Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.146692 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.147364 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.258116 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.416086 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2qdm\" (UniqueName: \"kubernetes.io/projected/9d368299-6a4f-4493-8460-3fc0e53f861c-kube-api-access-z2qdm\") pod \"9d368299-6a4f-4493-8460-3fc0e53f861c\" (UID: \"9d368299-6a4f-4493-8460-3fc0e53f861c\") " Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.416238 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d368299-6a4f-4493-8460-3fc0e53f861c-ssh-key\") pod \"9d368299-6a4f-4493-8460-3fc0e53f861c\" (UID: \"9d368299-6a4f-4493-8460-3fc0e53f861c\") " Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.416376 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d368299-6a4f-4493-8460-3fc0e53f861c-inventory\") pod \"9d368299-6a4f-4493-8460-3fc0e53f861c\" (UID: \"9d368299-6a4f-4493-8460-3fc0e53f861c\") " Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.421940 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d368299-6a4f-4493-8460-3fc0e53f861c-kube-api-access-z2qdm" (OuterVolumeSpecName: "kube-api-access-z2qdm") pod "9d368299-6a4f-4493-8460-3fc0e53f861c" (UID: "9d368299-6a4f-4493-8460-3fc0e53f861c"). InnerVolumeSpecName "kube-api-access-z2qdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.449041 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d368299-6a4f-4493-8460-3fc0e53f861c-inventory" (OuterVolumeSpecName: "inventory") pod "9d368299-6a4f-4493-8460-3fc0e53f861c" (UID: "9d368299-6a4f-4493-8460-3fc0e53f861c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.462505 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d368299-6a4f-4493-8460-3fc0e53f861c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9d368299-6a4f-4493-8460-3fc0e53f861c" (UID: "9d368299-6a4f-4493-8460-3fc0e53f861c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.518936 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2qdm\" (UniqueName: \"kubernetes.io/projected/9d368299-6a4f-4493-8460-3fc0e53f861c-kube-api-access-z2qdm\") on node \"crc\" DevicePath \"\"" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.518971 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d368299-6a4f-4493-8460-3fc0e53f861c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.518981 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d368299-6a4f-4493-8460-3fc0e53f861c-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.803060 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" event={"ID":"9d368299-6a4f-4493-8460-3fc0e53f861c","Type":"ContainerDied","Data":"17dbe2fdea10ec4c01edc5a813fc14872ee845bef88fee276bbb8e391ddf8771"} Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.803379 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17dbe2fdea10ec4c01edc5a813fc14872ee845bef88fee276bbb8e391ddf8771" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.803164 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.904762 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq"] Oct 02 10:13:42 crc kubenswrapper[4771]: E1002 10:13:42.905285 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d368299-6a4f-4493-8460-3fc0e53f861c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.905307 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d368299-6a4f-4493-8460-3fc0e53f861c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.905554 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d368299-6a4f-4493-8460-3fc0e53f861c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.906468 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.908236 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.909388 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.909530 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.909582 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.921186 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq"] Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.939773 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b873f5d-5502-4c9b-ba58-9eba9b62739b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq\" (UID: \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.939885 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b873f5d-5502-4c9b-ba58-9eba9b62739b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq\" (UID: \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" Oct 02 10:13:42 crc kubenswrapper[4771]: I1002 10:13:42.940100 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k89mw\" (UniqueName: \"kubernetes.io/projected/7b873f5d-5502-4c9b-ba58-9eba9b62739b-kube-api-access-k89mw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq\" (UID: \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" Oct 02 10:13:43 crc kubenswrapper[4771]: I1002 10:13:43.042188 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b873f5d-5502-4c9b-ba58-9eba9b62739b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq\" (UID: \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" Oct 02 10:13:43 crc kubenswrapper[4771]: I1002 10:13:43.042290 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b873f5d-5502-4c9b-ba58-9eba9b62739b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq\" (UID: \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" Oct 02 10:13:43 crc kubenswrapper[4771]: I1002 10:13:43.042385 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k89mw\" (UniqueName: \"kubernetes.io/projected/7b873f5d-5502-4c9b-ba58-9eba9b62739b-kube-api-access-k89mw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq\" (UID: \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" Oct 02 10:13:43 crc kubenswrapper[4771]: I1002 10:13:43.047275 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b873f5d-5502-4c9b-ba58-9eba9b62739b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq\" (UID: \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" Oct 02 10:13:43 crc kubenswrapper[4771]: I1002 10:13:43.055256 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b873f5d-5502-4c9b-ba58-9eba9b62739b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq\" (UID: \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" Oct 02 10:13:43 crc kubenswrapper[4771]: I1002 10:13:43.059106 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k89mw\" (UniqueName: \"kubernetes.io/projected/7b873f5d-5502-4c9b-ba58-9eba9b62739b-kube-api-access-k89mw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq\" (UID: \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" Oct 02 10:13:43 crc kubenswrapper[4771]: I1002 10:13:43.233773 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" Oct 02 10:13:43 crc kubenswrapper[4771]: I1002 10:13:43.761145 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq"] Oct 02 10:13:43 crc kubenswrapper[4771]: I1002 10:13:43.816199 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" event={"ID":"7b873f5d-5502-4c9b-ba58-9eba9b62739b","Type":"ContainerStarted","Data":"d876bdd3135a4cded38b69bb087eea0d82f398f669e8090daea32315d0cc71b4"} Oct 02 10:13:44 crc kubenswrapper[4771]: I1002 10:13:44.833450 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" event={"ID":"7b873f5d-5502-4c9b-ba58-9eba9b62739b","Type":"ContainerStarted","Data":"172476f429186ae90157ae9ba2d241a1fb3e83a8bcd81741d52c864ef4db32a8"} Oct 02 10:13:44 crc kubenswrapper[4771]: I1002 10:13:44.864098 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" podStartSLOduration=2.3957808 podStartE2EDuration="2.864051152s" podCreationTimestamp="2025-10-02 10:13:42 +0000 UTC" firstStartedPulling="2025-10-02 10:13:43.761494336 +0000 UTC m=+2211.409179403" lastFinishedPulling="2025-10-02 10:13:44.229764688 +0000 UTC m=+2211.877449755" observedRunningTime="2025-10-02 10:13:44.846684614 +0000 UTC m=+2212.494369701" watchObservedRunningTime="2025-10-02 10:13:44.864051152 +0000 UTC m=+2212.511736229" Oct 02 10:13:46 crc kubenswrapper[4771]: I1002 10:13:46.045367 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-mzcxs"] Oct 02 10:13:46 crc kubenswrapper[4771]: I1002 10:13:46.063720 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-djw9k"] Oct 02 10:13:46 crc kubenswrapper[4771]: I1002 10:13:46.076987 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-mzcxs"] Oct 02 10:13:46 crc kubenswrapper[4771]: I1002 10:13:46.085853 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-djw9k"] Oct 02 10:13:47 crc kubenswrapper[4771]: I1002 10:13:47.697566 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="729318ed-7618-49be-bcd1-43734b293975" path="/var/lib/kubelet/pods/729318ed-7618-49be-bcd1-43734b293975/volumes" Oct 02 10:13:47 crc kubenswrapper[4771]: I1002 10:13:47.699003 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="913531c6-8fa3-4c33-87a4-22f18bb251b8" path="/var/lib/kubelet/pods/913531c6-8fa3-4c33-87a4-22f18bb251b8/volumes" Oct 02 10:13:49 crc kubenswrapper[4771]: I1002 10:13:49.883964 4771 generic.go:334] "Generic (PLEG): container finished" podID="7b873f5d-5502-4c9b-ba58-9eba9b62739b" containerID="172476f429186ae90157ae9ba2d241a1fb3e83a8bcd81741d52c864ef4db32a8" exitCode=0 Oct 02 10:13:49 crc kubenswrapper[4771]: I1002 10:13:49.884060 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" event={"ID":"7b873f5d-5502-4c9b-ba58-9eba9b62739b","Type":"ContainerDied","Data":"172476f429186ae90157ae9ba2d241a1fb3e83a8bcd81741d52c864ef4db32a8"} Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.391824 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.482208 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k89mw\" (UniqueName: \"kubernetes.io/projected/7b873f5d-5502-4c9b-ba58-9eba9b62739b-kube-api-access-k89mw\") pod \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\" (UID: \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\") " Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.482341 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b873f5d-5502-4c9b-ba58-9eba9b62739b-ssh-key\") pod \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\" (UID: \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\") " Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.482443 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b873f5d-5502-4c9b-ba58-9eba9b62739b-inventory\") pod \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\" (UID: \"7b873f5d-5502-4c9b-ba58-9eba9b62739b\") " Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.488145 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b873f5d-5502-4c9b-ba58-9eba9b62739b-kube-api-access-k89mw" (OuterVolumeSpecName: "kube-api-access-k89mw") pod "7b873f5d-5502-4c9b-ba58-9eba9b62739b" (UID: "7b873f5d-5502-4c9b-ba58-9eba9b62739b"). InnerVolumeSpecName "kube-api-access-k89mw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.519678 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b873f5d-5502-4c9b-ba58-9eba9b62739b-inventory" (OuterVolumeSpecName: "inventory") pod "7b873f5d-5502-4c9b-ba58-9eba9b62739b" (UID: "7b873f5d-5502-4c9b-ba58-9eba9b62739b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.523062 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b873f5d-5502-4c9b-ba58-9eba9b62739b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7b873f5d-5502-4c9b-ba58-9eba9b62739b" (UID: "7b873f5d-5502-4c9b-ba58-9eba9b62739b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.585731 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k89mw\" (UniqueName: \"kubernetes.io/projected/7b873f5d-5502-4c9b-ba58-9eba9b62739b-kube-api-access-k89mw\") on node \"crc\" DevicePath \"\"" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.585768 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b873f5d-5502-4c9b-ba58-9eba9b62739b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.585777 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b873f5d-5502-4c9b-ba58-9eba9b62739b-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.907052 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" event={"ID":"7b873f5d-5502-4c9b-ba58-9eba9b62739b","Type":"ContainerDied","Data":"d876bdd3135a4cded38b69bb087eea0d82f398f669e8090daea32315d0cc71b4"} Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.907375 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d876bdd3135a4cded38b69bb087eea0d82f398f669e8090daea32315d0cc71b4" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.907094 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.979535 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v"] Oct 02 10:13:51 crc kubenswrapper[4771]: E1002 10:13:51.980409 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b873f5d-5502-4c9b-ba58-9eba9b62739b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.980500 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b873f5d-5502-4c9b-ba58-9eba9b62739b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.980957 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b873f5d-5502-4c9b-ba58-9eba9b62739b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.982004 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.987022 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.987069 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.987225 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.987238 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:13:51 crc kubenswrapper[4771]: I1002 10:13:51.990824 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v"] Oct 02 10:13:52 crc kubenswrapper[4771]: I1002 10:13:52.097191 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/224fd5aa-8697-424f-8960-dcf3a2462981-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hbq9v\" (UID: \"224fd5aa-8697-424f-8960-dcf3a2462981\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" Oct 02 10:13:52 crc kubenswrapper[4771]: I1002 10:13:52.097274 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xng86\" (UniqueName: \"kubernetes.io/projected/224fd5aa-8697-424f-8960-dcf3a2462981-kube-api-access-xng86\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hbq9v\" (UID: \"224fd5aa-8697-424f-8960-dcf3a2462981\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" Oct 02 10:13:52 crc kubenswrapper[4771]: I1002 10:13:52.097423 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/224fd5aa-8697-424f-8960-dcf3a2462981-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hbq9v\" (UID: \"224fd5aa-8697-424f-8960-dcf3a2462981\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" Oct 02 10:13:52 crc kubenswrapper[4771]: I1002 10:13:52.199540 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/224fd5aa-8697-424f-8960-dcf3a2462981-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hbq9v\" (UID: \"224fd5aa-8697-424f-8960-dcf3a2462981\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" Oct 02 10:13:52 crc kubenswrapper[4771]: I1002 10:13:52.199671 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/224fd5aa-8697-424f-8960-dcf3a2462981-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hbq9v\" (UID: \"224fd5aa-8697-424f-8960-dcf3a2462981\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" Oct 02 10:13:52 crc kubenswrapper[4771]: I1002 10:13:52.199722 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xng86\" (UniqueName: \"kubernetes.io/projected/224fd5aa-8697-424f-8960-dcf3a2462981-kube-api-access-xng86\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hbq9v\" (UID: \"224fd5aa-8697-424f-8960-dcf3a2462981\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" Oct 02 10:13:52 crc kubenswrapper[4771]: I1002 10:13:52.204549 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/224fd5aa-8697-424f-8960-dcf3a2462981-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hbq9v\" (UID: \"224fd5aa-8697-424f-8960-dcf3a2462981\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" Oct 02 10:13:52 crc kubenswrapper[4771]: I1002 10:13:52.209692 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/224fd5aa-8697-424f-8960-dcf3a2462981-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hbq9v\" (UID: \"224fd5aa-8697-424f-8960-dcf3a2462981\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" Oct 02 10:13:52 crc kubenswrapper[4771]: I1002 10:13:52.215856 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xng86\" (UniqueName: \"kubernetes.io/projected/224fd5aa-8697-424f-8960-dcf3a2462981-kube-api-access-xng86\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hbq9v\" (UID: \"224fd5aa-8697-424f-8960-dcf3a2462981\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" Oct 02 10:13:52 crc kubenswrapper[4771]: I1002 10:13:52.311807 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" Oct 02 10:13:52 crc kubenswrapper[4771]: I1002 10:13:52.815540 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v"] Oct 02 10:13:52 crc kubenswrapper[4771]: I1002 10:13:52.930657 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" event={"ID":"224fd5aa-8697-424f-8960-dcf3a2462981","Type":"ContainerStarted","Data":"de6653312aba5170daa8164e432d93aca17dea95821d141983097a5e1feec1a2"} Oct 02 10:13:53 crc kubenswrapper[4771]: I1002 10:13:53.943071 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" event={"ID":"224fd5aa-8697-424f-8960-dcf3a2462981","Type":"ContainerStarted","Data":"ed04102c9029f857929cd19470174c9bb8bb5269da5820c7b738e7ebba4e593f"} Oct 02 10:13:53 crc kubenswrapper[4771]: I1002 10:13:53.975239 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" podStartSLOduration=2.541410966 podStartE2EDuration="2.975219729s" podCreationTimestamp="2025-10-02 10:13:51 +0000 UTC" firstStartedPulling="2025-10-02 10:13:52.814272592 +0000 UTC m=+2220.461957659" lastFinishedPulling="2025-10-02 10:13:53.248081355 +0000 UTC m=+2220.895766422" observedRunningTime="2025-10-02 10:13:53.971189569 +0000 UTC m=+2221.618874636" watchObservedRunningTime="2025-10-02 10:13:53.975219729 +0000 UTC m=+2221.622904796" Oct 02 10:14:03 crc kubenswrapper[4771]: I1002 10:14:03.678056 4771 scope.go:117] "RemoveContainer" containerID="0ccc59478bfea768a2f882afbb3e5f6d0260e6d4a3d96c01bb215769d830ea0a" Oct 02 10:14:03 crc kubenswrapper[4771]: I1002 10:14:03.724073 4771 scope.go:117] "RemoveContainer" containerID="9165c06caa4386e66c7ead54f76dac3f12ac81c3c8c98c541908d959ccb9ea48" Oct 02 10:14:03 crc kubenswrapper[4771]: I1002 10:14:03.763156 4771 scope.go:117] "RemoveContainer" containerID="dc73f360304fbd06b30fb404084e0f2ef2aae0cdca71095d21d610bfb16c249e" Oct 02 10:14:03 crc kubenswrapper[4771]: I1002 10:14:03.814655 4771 scope.go:117] "RemoveContainer" containerID="cbfa7a7774deaaf3dfdb7596c8b2a9c83d804e29ba043e5befcae2aa740057db" Oct 02 10:14:03 crc kubenswrapper[4771]: I1002 10:14:03.866202 4771 scope.go:117] "RemoveContainer" containerID="557e4a57adaef22405d2d3b8213358bff7135decebaa687f4fa6039a1ad457f4" Oct 02 10:14:12 crc kubenswrapper[4771]: I1002 10:14:12.145863 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:14:12 crc kubenswrapper[4771]: I1002 10:14:12.146451 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:14:30 crc kubenswrapper[4771]: I1002 10:14:30.045626 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-96p7k"] Oct 02 10:14:30 crc kubenswrapper[4771]: I1002 10:14:30.057526 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-96p7k"] Oct 02 10:14:31 crc kubenswrapper[4771]: I1002 10:14:31.697245 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8da47e22-e6c2-4c9f-92f0-9ae49a82be69" path="/var/lib/kubelet/pods/8da47e22-e6c2-4c9f-92f0-9ae49a82be69/volumes" Oct 02 10:14:32 crc kubenswrapper[4771]: I1002 10:14:32.332771 4771 generic.go:334] "Generic (PLEG): container finished" podID="224fd5aa-8697-424f-8960-dcf3a2462981" containerID="ed04102c9029f857929cd19470174c9bb8bb5269da5820c7b738e7ebba4e593f" exitCode=0 Oct 02 10:14:32 crc kubenswrapper[4771]: I1002 10:14:32.332815 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" event={"ID":"224fd5aa-8697-424f-8960-dcf3a2462981","Type":"ContainerDied","Data":"ed04102c9029f857929cd19470174c9bb8bb5269da5820c7b738e7ebba4e593f"} Oct 02 10:14:33 crc kubenswrapper[4771]: I1002 10:14:33.823348 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" Oct 02 10:14:33 crc kubenswrapper[4771]: I1002 10:14:33.984610 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xng86\" (UniqueName: \"kubernetes.io/projected/224fd5aa-8697-424f-8960-dcf3a2462981-kube-api-access-xng86\") pod \"224fd5aa-8697-424f-8960-dcf3a2462981\" (UID: \"224fd5aa-8697-424f-8960-dcf3a2462981\") " Oct 02 10:14:33 crc kubenswrapper[4771]: I1002 10:14:33.984737 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/224fd5aa-8697-424f-8960-dcf3a2462981-inventory\") pod \"224fd5aa-8697-424f-8960-dcf3a2462981\" (UID: \"224fd5aa-8697-424f-8960-dcf3a2462981\") " Oct 02 10:14:33 crc kubenswrapper[4771]: I1002 10:14:33.985017 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/224fd5aa-8697-424f-8960-dcf3a2462981-ssh-key\") pod \"224fd5aa-8697-424f-8960-dcf3a2462981\" (UID: \"224fd5aa-8697-424f-8960-dcf3a2462981\") " Oct 02 10:14:33 crc kubenswrapper[4771]: I1002 10:14:33.990524 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/224fd5aa-8697-424f-8960-dcf3a2462981-kube-api-access-xng86" (OuterVolumeSpecName: "kube-api-access-xng86") pod "224fd5aa-8697-424f-8960-dcf3a2462981" (UID: "224fd5aa-8697-424f-8960-dcf3a2462981"). InnerVolumeSpecName "kube-api-access-xng86". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.017643 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/224fd5aa-8697-424f-8960-dcf3a2462981-inventory" (OuterVolumeSpecName: "inventory") pod "224fd5aa-8697-424f-8960-dcf3a2462981" (UID: "224fd5aa-8697-424f-8960-dcf3a2462981"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.050263 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/224fd5aa-8697-424f-8960-dcf3a2462981-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "224fd5aa-8697-424f-8960-dcf3a2462981" (UID: "224fd5aa-8697-424f-8960-dcf3a2462981"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.088118 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/224fd5aa-8697-424f-8960-dcf3a2462981-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.088168 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/224fd5aa-8697-424f-8960-dcf3a2462981-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.088183 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xng86\" (UniqueName: \"kubernetes.io/projected/224fd5aa-8697-424f-8960-dcf3a2462981-kube-api-access-xng86\") on node \"crc\" DevicePath \"\"" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.353592 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" event={"ID":"224fd5aa-8697-424f-8960-dcf3a2462981","Type":"ContainerDied","Data":"de6653312aba5170daa8164e432d93aca17dea95821d141983097a5e1feec1a2"} Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.353634 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hbq9v" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.353650 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de6653312aba5170daa8164e432d93aca17dea95821d141983097a5e1feec1a2" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.512917 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2"] Oct 02 10:14:34 crc kubenswrapper[4771]: E1002 10:14:34.513794 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224fd5aa-8697-424f-8960-dcf3a2462981" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.513821 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="224fd5aa-8697-424f-8960-dcf3a2462981" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.514395 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="224fd5aa-8697-424f-8960-dcf3a2462981" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.515596 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.518031 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.518985 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.519164 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.520358 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.527817 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2"] Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.700912 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a1795e1-9162-428a-8ebe-fcc5273eed05-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2\" (UID: \"5a1795e1-9162-428a-8ebe-fcc5273eed05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.701002 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a1795e1-9162-428a-8ebe-fcc5273eed05-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2\" (UID: \"5a1795e1-9162-428a-8ebe-fcc5273eed05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.701041 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z55c\" (UniqueName: \"kubernetes.io/projected/5a1795e1-9162-428a-8ebe-fcc5273eed05-kube-api-access-4z55c\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2\" (UID: \"5a1795e1-9162-428a-8ebe-fcc5273eed05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.803213 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a1795e1-9162-428a-8ebe-fcc5273eed05-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2\" (UID: \"5a1795e1-9162-428a-8ebe-fcc5273eed05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.803304 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a1795e1-9162-428a-8ebe-fcc5273eed05-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2\" (UID: \"5a1795e1-9162-428a-8ebe-fcc5273eed05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.803333 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z55c\" (UniqueName: \"kubernetes.io/projected/5a1795e1-9162-428a-8ebe-fcc5273eed05-kube-api-access-4z55c\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2\" (UID: \"5a1795e1-9162-428a-8ebe-fcc5273eed05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.807428 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a1795e1-9162-428a-8ebe-fcc5273eed05-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2\" (UID: \"5a1795e1-9162-428a-8ebe-fcc5273eed05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.811667 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a1795e1-9162-428a-8ebe-fcc5273eed05-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2\" (UID: \"5a1795e1-9162-428a-8ebe-fcc5273eed05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.824955 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z55c\" (UniqueName: \"kubernetes.io/projected/5a1795e1-9162-428a-8ebe-fcc5273eed05-kube-api-access-4z55c\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2\" (UID: \"5a1795e1-9162-428a-8ebe-fcc5273eed05\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" Oct 02 10:14:34 crc kubenswrapper[4771]: I1002 10:14:34.847303 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" Oct 02 10:14:35 crc kubenswrapper[4771]: I1002 10:14:35.461602 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2"] Oct 02 10:14:36 crc kubenswrapper[4771]: I1002 10:14:36.374632 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" event={"ID":"5a1795e1-9162-428a-8ebe-fcc5273eed05","Type":"ContainerStarted","Data":"c0c57668741e843ce3c9f65fc73646ef12a4e77b85e7c90db89c59ead7a81457"} Oct 02 10:14:36 crc kubenswrapper[4771]: I1002 10:14:36.375234 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" event={"ID":"5a1795e1-9162-428a-8ebe-fcc5273eed05","Type":"ContainerStarted","Data":"3977ca6ab1b3d86e6dd2f249341bb7d624859ead0feb0601934c3cca6c6d73f7"} Oct 02 10:14:36 crc kubenswrapper[4771]: I1002 10:14:36.408711 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" podStartSLOduration=1.871228142 podStartE2EDuration="2.408691398s" podCreationTimestamp="2025-10-02 10:14:34 +0000 UTC" firstStartedPulling="2025-10-02 10:14:35.447624265 +0000 UTC m=+2263.095309332" lastFinishedPulling="2025-10-02 10:14:35.985087521 +0000 UTC m=+2263.632772588" observedRunningTime="2025-10-02 10:14:36.402566966 +0000 UTC m=+2264.050252023" watchObservedRunningTime="2025-10-02 10:14:36.408691398 +0000 UTC m=+2264.056376465" Oct 02 10:14:42 crc kubenswrapper[4771]: I1002 10:14:42.145954 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:14:42 crc kubenswrapper[4771]: I1002 10:14:42.146529 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:14:42 crc kubenswrapper[4771]: I1002 10:14:42.146570 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 10:14:42 crc kubenswrapper[4771]: I1002 10:14:42.147515 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:14:42 crc kubenswrapper[4771]: I1002 10:14:42.147574 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" gracePeriod=600 Oct 02 10:14:42 crc kubenswrapper[4771]: E1002 10:14:42.268043 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:14:42 crc kubenswrapper[4771]: I1002 10:14:42.441045 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" exitCode=0 Oct 02 10:14:42 crc kubenswrapper[4771]: I1002 10:14:42.441095 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2"} Oct 02 10:14:42 crc kubenswrapper[4771]: I1002 10:14:42.441146 4771 scope.go:117] "RemoveContainer" containerID="ec7f5273f9ea2528fa747e0f9ebadd0f6e44b66e535ececeabf449a15093f50f" Oct 02 10:14:42 crc kubenswrapper[4771]: I1002 10:14:42.441980 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:14:42 crc kubenswrapper[4771]: E1002 10:14:42.442345 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:14:56 crc kubenswrapper[4771]: I1002 10:14:56.682545 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:14:56 crc kubenswrapper[4771]: E1002 10:14:56.684609 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.165372 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9"] Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.167625 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.181050 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.181745 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.184526 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9"] Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.349790 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-secret-volume\") pod \"collect-profiles-29323335-wt5h9\" (UID: \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.350082 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7422\" (UniqueName: \"kubernetes.io/projected/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-kube-api-access-f7422\") pod \"collect-profiles-29323335-wt5h9\" (UID: \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.350369 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-config-volume\") pod \"collect-profiles-29323335-wt5h9\" (UID: \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.452449 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7422\" (UniqueName: \"kubernetes.io/projected/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-kube-api-access-f7422\") pod \"collect-profiles-29323335-wt5h9\" (UID: \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.452570 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-config-volume\") pod \"collect-profiles-29323335-wt5h9\" (UID: \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.452745 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-secret-volume\") pod \"collect-profiles-29323335-wt5h9\" (UID: \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.453554 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-config-volume\") pod \"collect-profiles-29323335-wt5h9\" (UID: \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.461738 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-secret-volume\") pod \"collect-profiles-29323335-wt5h9\" (UID: \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.475272 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7422\" (UniqueName: \"kubernetes.io/projected/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-kube-api-access-f7422\") pod \"collect-profiles-29323335-wt5h9\" (UID: \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" Oct 02 10:15:00 crc kubenswrapper[4771]: I1002 10:15:00.527848 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" Oct 02 10:15:01 crc kubenswrapper[4771]: I1002 10:15:01.033331 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9"] Oct 02 10:15:01 crc kubenswrapper[4771]: I1002 10:15:01.666306 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" event={"ID":"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217","Type":"ContainerStarted","Data":"62529b0c27dd7e5451357bd026448129da580c3778f5fe72162f8c158870df52"} Oct 02 10:15:01 crc kubenswrapper[4771]: I1002 10:15:01.666674 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" event={"ID":"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217","Type":"ContainerStarted","Data":"70045d6b73dffcc2629eeb4abd152a5d52fb9fb506fdd6b28eeb5f796ff3a63a"} Oct 02 10:15:01 crc kubenswrapper[4771]: I1002 10:15:01.699685 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" podStartSLOduration=1.699662321 podStartE2EDuration="1.699662321s" podCreationTimestamp="2025-10-02 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:15:01.683280852 +0000 UTC m=+2289.330965919" watchObservedRunningTime="2025-10-02 10:15:01.699662321 +0000 UTC m=+2289.347347388" Oct 02 10:15:02 crc kubenswrapper[4771]: I1002 10:15:02.696320 4771 generic.go:334] "Generic (PLEG): container finished" podID="a9fc2c00-9a21-4c8e-ac5f-de3e660aa217" containerID="62529b0c27dd7e5451357bd026448129da580c3778f5fe72162f8c158870df52" exitCode=0 Oct 02 10:15:02 crc kubenswrapper[4771]: I1002 10:15:02.696472 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" event={"ID":"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217","Type":"ContainerDied","Data":"62529b0c27dd7e5451357bd026448129da580c3778f5fe72162f8c158870df52"} Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.030346 4771 scope.go:117] "RemoveContainer" containerID="0183f6869b55796a23e6c706865698dad8d57f152055b84e7b6a0e7d138391c6" Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.166219 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.266534 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-config-volume\") pod \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\" (UID: \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\") " Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.266587 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7422\" (UniqueName: \"kubernetes.io/projected/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-kube-api-access-f7422\") pod \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\" (UID: \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\") " Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.266635 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-secret-volume\") pod \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\" (UID: \"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217\") " Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.267543 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-config-volume" (OuterVolumeSpecName: "config-volume") pod "a9fc2c00-9a21-4c8e-ac5f-de3e660aa217" (UID: "a9fc2c00-9a21-4c8e-ac5f-de3e660aa217"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.272478 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a9fc2c00-9a21-4c8e-ac5f-de3e660aa217" (UID: "a9fc2c00-9a21-4c8e-ac5f-de3e660aa217"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.276424 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-kube-api-access-f7422" (OuterVolumeSpecName: "kube-api-access-f7422") pod "a9fc2c00-9a21-4c8e-ac5f-de3e660aa217" (UID: "a9fc2c00-9a21-4c8e-ac5f-de3e660aa217"). InnerVolumeSpecName "kube-api-access-f7422". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.371779 4771 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.371827 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7422\" (UniqueName: \"kubernetes.io/projected/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-kube-api-access-f7422\") on node \"crc\" DevicePath \"\"" Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.371849 4771 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.726394 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" event={"ID":"a9fc2c00-9a21-4c8e-ac5f-de3e660aa217","Type":"ContainerDied","Data":"70045d6b73dffcc2629eeb4abd152a5d52fb9fb506fdd6b28eeb5f796ff3a63a"} Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.726908 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70045d6b73dffcc2629eeb4abd152a5d52fb9fb506fdd6b28eeb5f796ff3a63a" Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.726946 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9" Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.761460 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r"] Oct 02 10:15:04 crc kubenswrapper[4771]: I1002 10:15:04.769966 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323290-clj5r"] Oct 02 10:15:05 crc kubenswrapper[4771]: I1002 10:15:05.701631 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e09f2d35-5bde-444c-866d-1376247d4659" path="/var/lib/kubelet/pods/e09f2d35-5bde-444c-866d-1376247d4659/volumes" Oct 02 10:15:11 crc kubenswrapper[4771]: I1002 10:15:11.683901 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:15:11 crc kubenswrapper[4771]: E1002 10:15:11.685085 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:15:26 crc kubenswrapper[4771]: I1002 10:15:26.682046 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:15:26 crc kubenswrapper[4771]: E1002 10:15:26.683359 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:15:35 crc kubenswrapper[4771]: I1002 10:15:35.076648 4771 generic.go:334] "Generic (PLEG): container finished" podID="5a1795e1-9162-428a-8ebe-fcc5273eed05" containerID="c0c57668741e843ce3c9f65fc73646ef12a4e77b85e7c90db89c59ead7a81457" exitCode=2 Oct 02 10:15:35 crc kubenswrapper[4771]: I1002 10:15:35.076735 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" event={"ID":"5a1795e1-9162-428a-8ebe-fcc5273eed05","Type":"ContainerDied","Data":"c0c57668741e843ce3c9f65fc73646ef12a4e77b85e7c90db89c59ead7a81457"} Oct 02 10:15:36 crc kubenswrapper[4771]: I1002 10:15:36.544281 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" Oct 02 10:15:36 crc kubenswrapper[4771]: I1002 10:15:36.682682 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z55c\" (UniqueName: \"kubernetes.io/projected/5a1795e1-9162-428a-8ebe-fcc5273eed05-kube-api-access-4z55c\") pod \"5a1795e1-9162-428a-8ebe-fcc5273eed05\" (UID: \"5a1795e1-9162-428a-8ebe-fcc5273eed05\") " Oct 02 10:15:36 crc kubenswrapper[4771]: I1002 10:15:36.683284 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a1795e1-9162-428a-8ebe-fcc5273eed05-inventory\") pod \"5a1795e1-9162-428a-8ebe-fcc5273eed05\" (UID: \"5a1795e1-9162-428a-8ebe-fcc5273eed05\") " Oct 02 10:15:36 crc kubenswrapper[4771]: I1002 10:15:36.683433 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a1795e1-9162-428a-8ebe-fcc5273eed05-ssh-key\") pod \"5a1795e1-9162-428a-8ebe-fcc5273eed05\" (UID: \"5a1795e1-9162-428a-8ebe-fcc5273eed05\") " Oct 02 10:15:36 crc kubenswrapper[4771]: I1002 10:15:36.705039 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a1795e1-9162-428a-8ebe-fcc5273eed05-kube-api-access-4z55c" (OuterVolumeSpecName: "kube-api-access-4z55c") pod "5a1795e1-9162-428a-8ebe-fcc5273eed05" (UID: "5a1795e1-9162-428a-8ebe-fcc5273eed05"). InnerVolumeSpecName "kube-api-access-4z55c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:15:36 crc kubenswrapper[4771]: I1002 10:15:36.718301 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1795e1-9162-428a-8ebe-fcc5273eed05-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5a1795e1-9162-428a-8ebe-fcc5273eed05" (UID: "5a1795e1-9162-428a-8ebe-fcc5273eed05"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:15:36 crc kubenswrapper[4771]: I1002 10:15:36.757018 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1795e1-9162-428a-8ebe-fcc5273eed05-inventory" (OuterVolumeSpecName: "inventory") pod "5a1795e1-9162-428a-8ebe-fcc5273eed05" (UID: "5a1795e1-9162-428a-8ebe-fcc5273eed05"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:15:36 crc kubenswrapper[4771]: I1002 10:15:36.787870 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a1795e1-9162-428a-8ebe-fcc5273eed05-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:15:36 crc kubenswrapper[4771]: I1002 10:15:36.787911 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4z55c\" (UniqueName: \"kubernetes.io/projected/5a1795e1-9162-428a-8ebe-fcc5273eed05-kube-api-access-4z55c\") on node \"crc\" DevicePath \"\"" Oct 02 10:15:36 crc kubenswrapper[4771]: I1002 10:15:36.787927 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a1795e1-9162-428a-8ebe-fcc5273eed05-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:15:37 crc kubenswrapper[4771]: I1002 10:15:37.104173 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" event={"ID":"5a1795e1-9162-428a-8ebe-fcc5273eed05","Type":"ContainerDied","Data":"3977ca6ab1b3d86e6dd2f249341bb7d624859ead0feb0601934c3cca6c6d73f7"} Oct 02 10:15:37 crc kubenswrapper[4771]: I1002 10:15:37.104240 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3977ca6ab1b3d86e6dd2f249341bb7d624859ead0feb0601934c3cca6c6d73f7" Oct 02 10:15:37 crc kubenswrapper[4771]: I1002 10:15:37.104244 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2" Oct 02 10:15:38 crc kubenswrapper[4771]: I1002 10:15:38.682008 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:15:38 crc kubenswrapper[4771]: E1002 10:15:38.682638 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.031351 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw"] Oct 02 10:15:44 crc kubenswrapper[4771]: E1002 10:15:44.032494 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a1795e1-9162-428a-8ebe-fcc5273eed05" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.032512 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a1795e1-9162-428a-8ebe-fcc5273eed05" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:15:44 crc kubenswrapper[4771]: E1002 10:15:44.032537 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9fc2c00-9a21-4c8e-ac5f-de3e660aa217" containerName="collect-profiles" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.032546 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9fc2c00-9a21-4c8e-ac5f-de3e660aa217" containerName="collect-profiles" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.032823 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a1795e1-9162-428a-8ebe-fcc5273eed05" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.032853 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9fc2c00-9a21-4c8e-ac5f-de3e660aa217" containerName="collect-profiles" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.033904 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.037444 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.037821 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.038078 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.039235 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.053892 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw"] Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.176838 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcgsh\" (UniqueName: \"kubernetes.io/projected/0842954b-9245-4437-be8d-9cae5cda66c4-kube-api-access-xcgsh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw\" (UID: \"0842954b-9245-4437-be8d-9cae5cda66c4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.177340 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0842954b-9245-4437-be8d-9cae5cda66c4-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw\" (UID: \"0842954b-9245-4437-be8d-9cae5cda66c4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.177812 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0842954b-9245-4437-be8d-9cae5cda66c4-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw\" (UID: \"0842954b-9245-4437-be8d-9cae5cda66c4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.280024 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0842954b-9245-4437-be8d-9cae5cda66c4-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw\" (UID: \"0842954b-9245-4437-be8d-9cae5cda66c4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.280210 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0842954b-9245-4437-be8d-9cae5cda66c4-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw\" (UID: \"0842954b-9245-4437-be8d-9cae5cda66c4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.280271 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcgsh\" (UniqueName: \"kubernetes.io/projected/0842954b-9245-4437-be8d-9cae5cda66c4-kube-api-access-xcgsh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw\" (UID: \"0842954b-9245-4437-be8d-9cae5cda66c4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.288994 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0842954b-9245-4437-be8d-9cae5cda66c4-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw\" (UID: \"0842954b-9245-4437-be8d-9cae5cda66c4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.289478 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0842954b-9245-4437-be8d-9cae5cda66c4-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw\" (UID: \"0842954b-9245-4437-be8d-9cae5cda66c4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.297345 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcgsh\" (UniqueName: \"kubernetes.io/projected/0842954b-9245-4437-be8d-9cae5cda66c4-kube-api-access-xcgsh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw\" (UID: \"0842954b-9245-4437-be8d-9cae5cda66c4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.360698 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.930378 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw"] Oct 02 10:15:44 crc kubenswrapper[4771]: I1002 10:15:44.940822 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:15:45 crc kubenswrapper[4771]: I1002 10:15:45.203226 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" event={"ID":"0842954b-9245-4437-be8d-9cae5cda66c4","Type":"ContainerStarted","Data":"fcf741f7fb8eefa954434acc6dbddbde719308d0dc26cb83646e53a9ad56a82f"} Oct 02 10:15:46 crc kubenswrapper[4771]: I1002 10:15:46.217771 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" event={"ID":"0842954b-9245-4437-be8d-9cae5cda66c4","Type":"ContainerStarted","Data":"6fbb6907468c67dd5e49b7116e6dc13259128ed18488921d9304d141f8731df0"} Oct 02 10:15:46 crc kubenswrapper[4771]: I1002 10:15:46.255966 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" podStartSLOduration=1.6084086690000001 podStartE2EDuration="2.255935775s" podCreationTimestamp="2025-10-02 10:15:44 +0000 UTC" firstStartedPulling="2025-10-02 10:15:44.940590845 +0000 UTC m=+2332.588275912" lastFinishedPulling="2025-10-02 10:15:45.588117951 +0000 UTC m=+2333.235803018" observedRunningTime="2025-10-02 10:15:46.243784979 +0000 UTC m=+2333.891470056" watchObservedRunningTime="2025-10-02 10:15:46.255935775 +0000 UTC m=+2333.903620872" Oct 02 10:15:51 crc kubenswrapper[4771]: I1002 10:15:51.682681 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:15:51 crc kubenswrapper[4771]: E1002 10:15:51.685897 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:16:04 crc kubenswrapper[4771]: I1002 10:16:04.183242 4771 scope.go:117] "RemoveContainer" containerID="633861a9f611c3118c907b03f0379cd2d5c5df9aa2226bd6fe9f56bb57798613" Oct 02 10:16:04 crc kubenswrapper[4771]: I1002 10:16:04.682105 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:16:04 crc kubenswrapper[4771]: E1002 10:16:04.682531 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:16:12 crc kubenswrapper[4771]: I1002 10:16:12.046717 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-h99xg"] Oct 02 10:16:12 crc kubenswrapper[4771]: I1002 10:16:12.064334 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-h99xg"] Oct 02 10:16:13 crc kubenswrapper[4771]: I1002 10:16:13.694069 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02" path="/var/lib/kubelet/pods/62d8bd0f-2fcf-4e0b-8268-ab1197e5ff02/volumes" Oct 02 10:16:19 crc kubenswrapper[4771]: I1002 10:16:19.681344 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:16:19 crc kubenswrapper[4771]: E1002 10:16:19.682305 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:16:20 crc kubenswrapper[4771]: I1002 10:16:20.529376 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c45z4"] Oct 02 10:16:20 crc kubenswrapper[4771]: I1002 10:16:20.532499 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:20 crc kubenswrapper[4771]: I1002 10:16:20.546369 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c45z4"] Oct 02 10:16:20 crc kubenswrapper[4771]: I1002 10:16:20.628782 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wnsp\" (UniqueName: \"kubernetes.io/projected/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-kube-api-access-5wnsp\") pod \"redhat-marketplace-c45z4\" (UID: \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\") " pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:20 crc kubenswrapper[4771]: I1002 10:16:20.628856 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-utilities\") pod \"redhat-marketplace-c45z4\" (UID: \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\") " pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:20 crc kubenswrapper[4771]: I1002 10:16:20.628892 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-catalog-content\") pod \"redhat-marketplace-c45z4\" (UID: \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\") " pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:20 crc kubenswrapper[4771]: I1002 10:16:20.731409 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wnsp\" (UniqueName: \"kubernetes.io/projected/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-kube-api-access-5wnsp\") pod \"redhat-marketplace-c45z4\" (UID: \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\") " pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:20 crc kubenswrapper[4771]: I1002 10:16:20.731776 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-utilities\") pod \"redhat-marketplace-c45z4\" (UID: \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\") " pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:20 crc kubenswrapper[4771]: I1002 10:16:20.731803 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-catalog-content\") pod \"redhat-marketplace-c45z4\" (UID: \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\") " pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:20 crc kubenswrapper[4771]: I1002 10:16:20.732250 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-utilities\") pod \"redhat-marketplace-c45z4\" (UID: \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\") " pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:20 crc kubenswrapper[4771]: I1002 10:16:20.732411 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-catalog-content\") pod \"redhat-marketplace-c45z4\" (UID: \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\") " pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:20 crc kubenswrapper[4771]: I1002 10:16:20.749938 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wnsp\" (UniqueName: \"kubernetes.io/projected/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-kube-api-access-5wnsp\") pod \"redhat-marketplace-c45z4\" (UID: \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\") " pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:20 crc kubenswrapper[4771]: I1002 10:16:20.859441 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:21 crc kubenswrapper[4771]: I1002 10:16:21.442392 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c45z4"] Oct 02 10:16:21 crc kubenswrapper[4771]: I1002 10:16:21.612641 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c45z4" event={"ID":"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3","Type":"ContainerStarted","Data":"f1b7e336d68b23ccfce42dcf57a2873299da0c504718289ba52e3fd41eb4e1a0"} Oct 02 10:16:22 crc kubenswrapper[4771]: I1002 10:16:22.625019 4771 generic.go:334] "Generic (PLEG): container finished" podID="f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" containerID="ac63d8f8bd08a37137280a1e0c475c817f6325f63bf82537c4b6d1cc26c8bc07" exitCode=0 Oct 02 10:16:22 crc kubenswrapper[4771]: I1002 10:16:22.625172 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c45z4" event={"ID":"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3","Type":"ContainerDied","Data":"ac63d8f8bd08a37137280a1e0c475c817f6325f63bf82537c4b6d1cc26c8bc07"} Oct 02 10:16:23 crc kubenswrapper[4771]: I1002 10:16:23.637764 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c45z4" event={"ID":"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3","Type":"ContainerStarted","Data":"574b6865f05a477b36a7fc54b893e5bb067158fd4d6d50b378fababae173d58f"} Oct 02 10:16:24 crc kubenswrapper[4771]: I1002 10:16:24.650709 4771 generic.go:334] "Generic (PLEG): container finished" podID="f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" containerID="574b6865f05a477b36a7fc54b893e5bb067158fd4d6d50b378fababae173d58f" exitCode=0 Oct 02 10:16:24 crc kubenswrapper[4771]: I1002 10:16:24.650754 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c45z4" event={"ID":"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3","Type":"ContainerDied","Data":"574b6865f05a477b36a7fc54b893e5bb067158fd4d6d50b378fababae173d58f"} Oct 02 10:16:25 crc kubenswrapper[4771]: I1002 10:16:25.664559 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c45z4" event={"ID":"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3","Type":"ContainerStarted","Data":"912059ab412db2570bafedcfe284a4fb43a9c2cbbf7f6cca6c7ccaed32a86e02"} Oct 02 10:16:25 crc kubenswrapper[4771]: I1002 10:16:25.710181 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c45z4" podStartSLOduration=3.205782372 podStartE2EDuration="5.710153227s" podCreationTimestamp="2025-10-02 10:16:20 +0000 UTC" firstStartedPulling="2025-10-02 10:16:22.627872407 +0000 UTC m=+2370.275557484" lastFinishedPulling="2025-10-02 10:16:25.132243272 +0000 UTC m=+2372.779928339" observedRunningTime="2025-10-02 10:16:25.694071079 +0000 UTC m=+2373.341756136" watchObservedRunningTime="2025-10-02 10:16:25.710153227 +0000 UTC m=+2373.357838294" Oct 02 10:16:30 crc kubenswrapper[4771]: I1002 10:16:30.860523 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:30 crc kubenswrapper[4771]: I1002 10:16:30.861097 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:30 crc kubenswrapper[4771]: I1002 10:16:30.929555 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:31 crc kubenswrapper[4771]: I1002 10:16:31.822898 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:31 crc kubenswrapper[4771]: I1002 10:16:31.883996 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c45z4"] Oct 02 10:16:32 crc kubenswrapper[4771]: I1002 10:16:32.683930 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:16:32 crc kubenswrapper[4771]: E1002 10:16:32.684472 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:16:33 crc kubenswrapper[4771]: I1002 10:16:33.770044 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c45z4" podUID="f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" containerName="registry-server" containerID="cri-o://912059ab412db2570bafedcfe284a4fb43a9c2cbbf7f6cca6c7ccaed32a86e02" gracePeriod=2 Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.221426 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.367035 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-utilities\") pod \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\" (UID: \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\") " Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.367158 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wnsp\" (UniqueName: \"kubernetes.io/projected/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-kube-api-access-5wnsp\") pod \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\" (UID: \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\") " Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.367243 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-catalog-content\") pod \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\" (UID: \"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3\") " Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.368418 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-utilities" (OuterVolumeSpecName: "utilities") pod "f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" (UID: "f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.373061 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-kube-api-access-5wnsp" (OuterVolumeSpecName: "kube-api-access-5wnsp") pod "f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" (UID: "f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3"). InnerVolumeSpecName "kube-api-access-5wnsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.379818 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" (UID: "f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.470067 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.470111 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wnsp\" (UniqueName: \"kubernetes.io/projected/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-kube-api-access-5wnsp\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.470143 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.784656 4771 generic.go:334] "Generic (PLEG): container finished" podID="f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" containerID="912059ab412db2570bafedcfe284a4fb43a9c2cbbf7f6cca6c7ccaed32a86e02" exitCode=0 Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.784702 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c45z4" event={"ID":"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3","Type":"ContainerDied","Data":"912059ab412db2570bafedcfe284a4fb43a9c2cbbf7f6cca6c7ccaed32a86e02"} Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.784723 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c45z4" Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.784761 4771 scope.go:117] "RemoveContainer" containerID="912059ab412db2570bafedcfe284a4fb43a9c2cbbf7f6cca6c7ccaed32a86e02" Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.784744 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c45z4" event={"ID":"f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3","Type":"ContainerDied","Data":"f1b7e336d68b23ccfce42dcf57a2873299da0c504718289ba52e3fd41eb4e1a0"} Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.823630 4771 scope.go:117] "RemoveContainer" containerID="574b6865f05a477b36a7fc54b893e5bb067158fd4d6d50b378fababae173d58f" Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.826248 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c45z4"] Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.846431 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c45z4"] Oct 02 10:16:34 crc kubenswrapper[4771]: I1002 10:16:34.878462 4771 scope.go:117] "RemoveContainer" containerID="ac63d8f8bd08a37137280a1e0c475c817f6325f63bf82537c4b6d1cc26c8bc07" Oct 02 10:16:35 crc kubenswrapper[4771]: I1002 10:16:35.005394 4771 scope.go:117] "RemoveContainer" containerID="912059ab412db2570bafedcfe284a4fb43a9c2cbbf7f6cca6c7ccaed32a86e02" Oct 02 10:16:35 crc kubenswrapper[4771]: E1002 10:16:35.008650 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"912059ab412db2570bafedcfe284a4fb43a9c2cbbf7f6cca6c7ccaed32a86e02\": container with ID starting with 912059ab412db2570bafedcfe284a4fb43a9c2cbbf7f6cca6c7ccaed32a86e02 not found: ID does not exist" containerID="912059ab412db2570bafedcfe284a4fb43a9c2cbbf7f6cca6c7ccaed32a86e02" Oct 02 10:16:35 crc kubenswrapper[4771]: I1002 10:16:35.008704 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"912059ab412db2570bafedcfe284a4fb43a9c2cbbf7f6cca6c7ccaed32a86e02"} err="failed to get container status \"912059ab412db2570bafedcfe284a4fb43a9c2cbbf7f6cca6c7ccaed32a86e02\": rpc error: code = NotFound desc = could not find container \"912059ab412db2570bafedcfe284a4fb43a9c2cbbf7f6cca6c7ccaed32a86e02\": container with ID starting with 912059ab412db2570bafedcfe284a4fb43a9c2cbbf7f6cca6c7ccaed32a86e02 not found: ID does not exist" Oct 02 10:16:35 crc kubenswrapper[4771]: I1002 10:16:35.008732 4771 scope.go:117] "RemoveContainer" containerID="574b6865f05a477b36a7fc54b893e5bb067158fd4d6d50b378fababae173d58f" Oct 02 10:16:35 crc kubenswrapper[4771]: E1002 10:16:35.012616 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"574b6865f05a477b36a7fc54b893e5bb067158fd4d6d50b378fababae173d58f\": container with ID starting with 574b6865f05a477b36a7fc54b893e5bb067158fd4d6d50b378fababae173d58f not found: ID does not exist" containerID="574b6865f05a477b36a7fc54b893e5bb067158fd4d6d50b378fababae173d58f" Oct 02 10:16:35 crc kubenswrapper[4771]: I1002 10:16:35.012676 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"574b6865f05a477b36a7fc54b893e5bb067158fd4d6d50b378fababae173d58f"} err="failed to get container status \"574b6865f05a477b36a7fc54b893e5bb067158fd4d6d50b378fababae173d58f\": rpc error: code = NotFound desc = could not find container \"574b6865f05a477b36a7fc54b893e5bb067158fd4d6d50b378fababae173d58f\": container with ID starting with 574b6865f05a477b36a7fc54b893e5bb067158fd4d6d50b378fababae173d58f not found: ID does not exist" Oct 02 10:16:35 crc kubenswrapper[4771]: I1002 10:16:35.012712 4771 scope.go:117] "RemoveContainer" containerID="ac63d8f8bd08a37137280a1e0c475c817f6325f63bf82537c4b6d1cc26c8bc07" Oct 02 10:16:35 crc kubenswrapper[4771]: E1002 10:16:35.013356 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac63d8f8bd08a37137280a1e0c475c817f6325f63bf82537c4b6d1cc26c8bc07\": container with ID starting with ac63d8f8bd08a37137280a1e0c475c817f6325f63bf82537c4b6d1cc26c8bc07 not found: ID does not exist" containerID="ac63d8f8bd08a37137280a1e0c475c817f6325f63bf82537c4b6d1cc26c8bc07" Oct 02 10:16:35 crc kubenswrapper[4771]: I1002 10:16:35.013394 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac63d8f8bd08a37137280a1e0c475c817f6325f63bf82537c4b6d1cc26c8bc07"} err="failed to get container status \"ac63d8f8bd08a37137280a1e0c475c817f6325f63bf82537c4b6d1cc26c8bc07\": rpc error: code = NotFound desc = could not find container \"ac63d8f8bd08a37137280a1e0c475c817f6325f63bf82537c4b6d1cc26c8bc07\": container with ID starting with ac63d8f8bd08a37137280a1e0c475c817f6325f63bf82537c4b6d1cc26c8bc07 not found: ID does not exist" Oct 02 10:16:35 crc kubenswrapper[4771]: I1002 10:16:35.705282 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" path="/var/lib/kubelet/pods/f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3/volumes" Oct 02 10:16:36 crc kubenswrapper[4771]: I1002 10:16:36.813456 4771 generic.go:334] "Generic (PLEG): container finished" podID="0842954b-9245-4437-be8d-9cae5cda66c4" containerID="6fbb6907468c67dd5e49b7116e6dc13259128ed18488921d9304d141f8731df0" exitCode=0 Oct 02 10:16:36 crc kubenswrapper[4771]: I1002 10:16:36.813565 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" event={"ID":"0842954b-9245-4437-be8d-9cae5cda66c4","Type":"ContainerDied","Data":"6fbb6907468c67dd5e49b7116e6dc13259128ed18488921d9304d141f8731df0"} Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.313786 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.476377 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0842954b-9245-4437-be8d-9cae5cda66c4-ssh-key\") pod \"0842954b-9245-4437-be8d-9cae5cda66c4\" (UID: \"0842954b-9245-4437-be8d-9cae5cda66c4\") " Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.476826 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0842954b-9245-4437-be8d-9cae5cda66c4-inventory\") pod \"0842954b-9245-4437-be8d-9cae5cda66c4\" (UID: \"0842954b-9245-4437-be8d-9cae5cda66c4\") " Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.476865 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgsh\" (UniqueName: \"kubernetes.io/projected/0842954b-9245-4437-be8d-9cae5cda66c4-kube-api-access-xcgsh\") pod \"0842954b-9245-4437-be8d-9cae5cda66c4\" (UID: \"0842954b-9245-4437-be8d-9cae5cda66c4\") " Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.486578 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0842954b-9245-4437-be8d-9cae5cda66c4-kube-api-access-xcgsh" (OuterVolumeSpecName: "kube-api-access-xcgsh") pod "0842954b-9245-4437-be8d-9cae5cda66c4" (UID: "0842954b-9245-4437-be8d-9cae5cda66c4"). InnerVolumeSpecName "kube-api-access-xcgsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.507985 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0842954b-9245-4437-be8d-9cae5cda66c4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0842954b-9245-4437-be8d-9cae5cda66c4" (UID: "0842954b-9245-4437-be8d-9cae5cda66c4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.510004 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0842954b-9245-4437-be8d-9cae5cda66c4-inventory" (OuterVolumeSpecName: "inventory") pod "0842954b-9245-4437-be8d-9cae5cda66c4" (UID: "0842954b-9245-4437-be8d-9cae5cda66c4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.580050 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0842954b-9245-4437-be8d-9cae5cda66c4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.580096 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0842954b-9245-4437-be8d-9cae5cda66c4-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.580111 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgsh\" (UniqueName: \"kubernetes.io/projected/0842954b-9245-4437-be8d-9cae5cda66c4-kube-api-access-xcgsh\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.841240 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" event={"ID":"0842954b-9245-4437-be8d-9cae5cda66c4","Type":"ContainerDied","Data":"fcf741f7fb8eefa954434acc6dbddbde719308d0dc26cb83646e53a9ad56a82f"} Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.841293 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcf741f7fb8eefa954434acc6dbddbde719308d0dc26cb83646e53a9ad56a82f" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.841327 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.931076 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5g9vt"] Oct 02 10:16:38 crc kubenswrapper[4771]: E1002 10:16:38.931618 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0842954b-9245-4437-be8d-9cae5cda66c4" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.931642 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="0842954b-9245-4437-be8d-9cae5cda66c4" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:16:38 crc kubenswrapper[4771]: E1002 10:16:38.931683 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" containerName="registry-server" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.931690 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" containerName="registry-server" Oct 02 10:16:38 crc kubenswrapper[4771]: E1002 10:16:38.931705 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" containerName="extract-utilities" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.931711 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" containerName="extract-utilities" Oct 02 10:16:38 crc kubenswrapper[4771]: E1002 10:16:38.931724 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" containerName="extract-content" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.931730 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" containerName="extract-content" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.931955 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1123b5c-f0d1-465a-a7f1-cb6b48f0b3c3" containerName="registry-server" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.931970 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="0842954b-9245-4437-be8d-9cae5cda66c4" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.932840 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.935346 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.935413 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.935623 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.935630 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:16:38 crc kubenswrapper[4771]: I1002 10:16:38.949608 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5g9vt"] Oct 02 10:16:39 crc kubenswrapper[4771]: I1002 10:16:39.090800 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqd48\" (UniqueName: \"kubernetes.io/projected/2315a9df-8051-4ec6-98a2-d8e0e32f598b-kube-api-access-gqd48\") pod \"ssh-known-hosts-edpm-deployment-5g9vt\" (UID: \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" Oct 02 10:16:39 crc kubenswrapper[4771]: I1002 10:16:39.090842 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2315a9df-8051-4ec6-98a2-d8e0e32f598b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5g9vt\" (UID: \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" Oct 02 10:16:39 crc kubenswrapper[4771]: I1002 10:16:39.090906 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2315a9df-8051-4ec6-98a2-d8e0e32f598b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5g9vt\" (UID: \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" Oct 02 10:16:39 crc kubenswrapper[4771]: I1002 10:16:39.193343 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqd48\" (UniqueName: \"kubernetes.io/projected/2315a9df-8051-4ec6-98a2-d8e0e32f598b-kube-api-access-gqd48\") pod \"ssh-known-hosts-edpm-deployment-5g9vt\" (UID: \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" Oct 02 10:16:39 crc kubenswrapper[4771]: I1002 10:16:39.193394 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2315a9df-8051-4ec6-98a2-d8e0e32f598b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5g9vt\" (UID: \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" Oct 02 10:16:39 crc kubenswrapper[4771]: I1002 10:16:39.193463 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2315a9df-8051-4ec6-98a2-d8e0e32f598b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5g9vt\" (UID: \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" Oct 02 10:16:39 crc kubenswrapper[4771]: I1002 10:16:39.198750 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2315a9df-8051-4ec6-98a2-d8e0e32f598b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5g9vt\" (UID: \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" Oct 02 10:16:39 crc kubenswrapper[4771]: I1002 10:16:39.198766 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2315a9df-8051-4ec6-98a2-d8e0e32f598b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5g9vt\" (UID: \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" Oct 02 10:16:39 crc kubenswrapper[4771]: I1002 10:16:39.227555 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqd48\" (UniqueName: \"kubernetes.io/projected/2315a9df-8051-4ec6-98a2-d8e0e32f598b-kube-api-access-gqd48\") pod \"ssh-known-hosts-edpm-deployment-5g9vt\" (UID: \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\") " pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" Oct 02 10:16:39 crc kubenswrapper[4771]: I1002 10:16:39.250664 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" Oct 02 10:16:39 crc kubenswrapper[4771]: I1002 10:16:39.845787 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5g9vt"] Oct 02 10:16:40 crc kubenswrapper[4771]: I1002 10:16:40.871996 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" event={"ID":"2315a9df-8051-4ec6-98a2-d8e0e32f598b","Type":"ContainerStarted","Data":"2b3197fadfb7881da1ad92affeb091937545992f52b16e77305dc3b8e4dd0ecc"} Oct 02 10:16:40 crc kubenswrapper[4771]: I1002 10:16:40.872604 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" event={"ID":"2315a9df-8051-4ec6-98a2-d8e0e32f598b","Type":"ContainerStarted","Data":"b54e674676e0a0b16f418e07bf184b82b680186772fce7b5967eaca906a6a644"} Oct 02 10:16:40 crc kubenswrapper[4771]: I1002 10:16:40.889767 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" podStartSLOduration=2.446631911 podStartE2EDuration="2.889749226s" podCreationTimestamp="2025-10-02 10:16:38 +0000 UTC" firstStartedPulling="2025-10-02 10:16:39.858985592 +0000 UTC m=+2387.506670679" lastFinishedPulling="2025-10-02 10:16:40.302102927 +0000 UTC m=+2387.949787994" observedRunningTime="2025-10-02 10:16:40.889224862 +0000 UTC m=+2388.536909949" watchObservedRunningTime="2025-10-02 10:16:40.889749226 +0000 UTC m=+2388.537434293" Oct 02 10:16:45 crc kubenswrapper[4771]: I1002 10:16:45.682040 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:16:45 crc kubenswrapper[4771]: E1002 10:16:45.682906 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:16:48 crc kubenswrapper[4771]: I1002 10:16:48.953582 4771 generic.go:334] "Generic (PLEG): container finished" podID="2315a9df-8051-4ec6-98a2-d8e0e32f598b" containerID="2b3197fadfb7881da1ad92affeb091937545992f52b16e77305dc3b8e4dd0ecc" exitCode=0 Oct 02 10:16:48 crc kubenswrapper[4771]: I1002 10:16:48.954181 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" event={"ID":"2315a9df-8051-4ec6-98a2-d8e0e32f598b","Type":"ContainerDied","Data":"2b3197fadfb7881da1ad92affeb091937545992f52b16e77305dc3b8e4dd0ecc"} Oct 02 10:16:49 crc kubenswrapper[4771]: I1002 10:16:49.049411 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-l5pd4"] Oct 02 10:16:49 crc kubenswrapper[4771]: I1002 10:16:49.062316 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-l5pd4"] Oct 02 10:16:49 crc kubenswrapper[4771]: I1002 10:16:49.697481 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94e19f78-6425-4099-8188-8dabe4c2d9f0" path="/var/lib/kubelet/pods/94e19f78-6425-4099-8188-8dabe4c2d9f0/volumes" Oct 02 10:16:50 crc kubenswrapper[4771]: I1002 10:16:50.448946 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" Oct 02 10:16:50 crc kubenswrapper[4771]: I1002 10:16:50.563604 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2315a9df-8051-4ec6-98a2-d8e0e32f598b-inventory-0\") pod \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\" (UID: \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\") " Oct 02 10:16:50 crc kubenswrapper[4771]: I1002 10:16:50.563669 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqd48\" (UniqueName: \"kubernetes.io/projected/2315a9df-8051-4ec6-98a2-d8e0e32f598b-kube-api-access-gqd48\") pod \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\" (UID: \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\") " Oct 02 10:16:50 crc kubenswrapper[4771]: I1002 10:16:50.563729 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2315a9df-8051-4ec6-98a2-d8e0e32f598b-ssh-key-openstack-edpm-ipam\") pod \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\" (UID: \"2315a9df-8051-4ec6-98a2-d8e0e32f598b\") " Oct 02 10:16:50 crc kubenswrapper[4771]: I1002 10:16:50.569485 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2315a9df-8051-4ec6-98a2-d8e0e32f598b-kube-api-access-gqd48" (OuterVolumeSpecName: "kube-api-access-gqd48") pod "2315a9df-8051-4ec6-98a2-d8e0e32f598b" (UID: "2315a9df-8051-4ec6-98a2-d8e0e32f598b"). InnerVolumeSpecName "kube-api-access-gqd48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:16:50 crc kubenswrapper[4771]: I1002 10:16:50.594643 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2315a9df-8051-4ec6-98a2-d8e0e32f598b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2315a9df-8051-4ec6-98a2-d8e0e32f598b" (UID: "2315a9df-8051-4ec6-98a2-d8e0e32f598b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:16:50 crc kubenswrapper[4771]: I1002 10:16:50.598765 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2315a9df-8051-4ec6-98a2-d8e0e32f598b-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "2315a9df-8051-4ec6-98a2-d8e0e32f598b" (UID: "2315a9df-8051-4ec6-98a2-d8e0e32f598b"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:16:50 crc kubenswrapper[4771]: I1002 10:16:50.667179 4771 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2315a9df-8051-4ec6-98a2-d8e0e32f598b-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:50 crc kubenswrapper[4771]: I1002 10:16:50.667208 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqd48\" (UniqueName: \"kubernetes.io/projected/2315a9df-8051-4ec6-98a2-d8e0e32f598b-kube-api-access-gqd48\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:50 crc kubenswrapper[4771]: I1002 10:16:50.667219 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2315a9df-8051-4ec6-98a2-d8e0e32f598b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:50 crc kubenswrapper[4771]: I1002 10:16:50.975825 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" event={"ID":"2315a9df-8051-4ec6-98a2-d8e0e32f598b","Type":"ContainerDied","Data":"b54e674676e0a0b16f418e07bf184b82b680186772fce7b5967eaca906a6a644"} Oct 02 10:16:50 crc kubenswrapper[4771]: I1002 10:16:50.975868 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b54e674676e0a0b16f418e07bf184b82b680186772fce7b5967eaca906a6a644" Oct 02 10:16:50 crc kubenswrapper[4771]: I1002 10:16:50.975918 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5g9vt" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.067021 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t"] Oct 02 10:16:51 crc kubenswrapper[4771]: E1002 10:16:51.067629 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2315a9df-8051-4ec6-98a2-d8e0e32f598b" containerName="ssh-known-hosts-edpm-deployment" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.067673 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2315a9df-8051-4ec6-98a2-d8e0e32f598b" containerName="ssh-known-hosts-edpm-deployment" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.068000 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2315a9df-8051-4ec6-98a2-d8e0e32f598b" containerName="ssh-known-hosts-edpm-deployment" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.068936 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.072461 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.072621 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.072870 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.073424 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.083024 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t"] Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.179723 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb23df5-8bb9-41fa-999e-6a563575b760-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dsv2t\" (UID: \"2cb23df5-8bb9-41fa-999e-6a563575b760\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.179848 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxnbz\" (UniqueName: \"kubernetes.io/projected/2cb23df5-8bb9-41fa-999e-6a563575b760-kube-api-access-lxnbz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dsv2t\" (UID: \"2cb23df5-8bb9-41fa-999e-6a563575b760\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.180261 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb23df5-8bb9-41fa-999e-6a563575b760-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dsv2t\" (UID: \"2cb23df5-8bb9-41fa-999e-6a563575b760\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.285087 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxnbz\" (UniqueName: \"kubernetes.io/projected/2cb23df5-8bb9-41fa-999e-6a563575b760-kube-api-access-lxnbz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dsv2t\" (UID: \"2cb23df5-8bb9-41fa-999e-6a563575b760\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.285334 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb23df5-8bb9-41fa-999e-6a563575b760-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dsv2t\" (UID: \"2cb23df5-8bb9-41fa-999e-6a563575b760\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.285397 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb23df5-8bb9-41fa-999e-6a563575b760-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dsv2t\" (UID: \"2cb23df5-8bb9-41fa-999e-6a563575b760\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.289403 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb23df5-8bb9-41fa-999e-6a563575b760-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dsv2t\" (UID: \"2cb23df5-8bb9-41fa-999e-6a563575b760\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.289534 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb23df5-8bb9-41fa-999e-6a563575b760-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dsv2t\" (UID: \"2cb23df5-8bb9-41fa-999e-6a563575b760\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.301512 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxnbz\" (UniqueName: \"kubernetes.io/projected/2cb23df5-8bb9-41fa-999e-6a563575b760-kube-api-access-lxnbz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dsv2t\" (UID: \"2cb23df5-8bb9-41fa-999e-6a563575b760\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" Oct 02 10:16:51 crc kubenswrapper[4771]: I1002 10:16:51.394029 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" Oct 02 10:16:52 crc kubenswrapper[4771]: I1002 10:16:52.023794 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t"] Oct 02 10:16:53 crc kubenswrapper[4771]: I1002 10:16:53.023371 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" event={"ID":"2cb23df5-8bb9-41fa-999e-6a563575b760","Type":"ContainerStarted","Data":"900a16fb078ab24e8b1caafd705ceb50d4634139393892abaacd057d99e427cb"} Oct 02 10:16:53 crc kubenswrapper[4771]: I1002 10:16:53.023947 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" event={"ID":"2cb23df5-8bb9-41fa-999e-6a563575b760","Type":"ContainerStarted","Data":"cc3716856b8a945c624cf5b97e8ef2f101bf709dfe8ca891719d5cafd42bb76b"} Oct 02 10:16:53 crc kubenswrapper[4771]: I1002 10:16:53.051815 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" podStartSLOduration=1.42144497 podStartE2EDuration="2.051787489s" podCreationTimestamp="2025-10-02 10:16:51 +0000 UTC" firstStartedPulling="2025-10-02 10:16:52.022382431 +0000 UTC m=+2399.670067498" lastFinishedPulling="2025-10-02 10:16:52.65272495 +0000 UTC m=+2400.300410017" observedRunningTime="2025-10-02 10:16:53.04146896 +0000 UTC m=+2400.689154027" watchObservedRunningTime="2025-10-02 10:16:53.051787489 +0000 UTC m=+2400.699472566" Oct 02 10:17:00 crc kubenswrapper[4771]: I1002 10:17:00.681655 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:17:00 crc kubenswrapper[4771]: E1002 10:17:00.682483 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:17:02 crc kubenswrapper[4771]: I1002 10:17:02.133256 4771 generic.go:334] "Generic (PLEG): container finished" podID="2cb23df5-8bb9-41fa-999e-6a563575b760" containerID="900a16fb078ab24e8b1caafd705ceb50d4634139393892abaacd057d99e427cb" exitCode=0 Oct 02 10:17:02 crc kubenswrapper[4771]: I1002 10:17:02.133324 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" event={"ID":"2cb23df5-8bb9-41fa-999e-6a563575b760","Type":"ContainerDied","Data":"900a16fb078ab24e8b1caafd705ceb50d4634139393892abaacd057d99e427cb"} Oct 02 10:17:03 crc kubenswrapper[4771]: I1002 10:17:03.620588 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" Oct 02 10:17:03 crc kubenswrapper[4771]: I1002 10:17:03.703698 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb23df5-8bb9-41fa-999e-6a563575b760-inventory\") pod \"2cb23df5-8bb9-41fa-999e-6a563575b760\" (UID: \"2cb23df5-8bb9-41fa-999e-6a563575b760\") " Oct 02 10:17:03 crc kubenswrapper[4771]: I1002 10:17:03.703896 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxnbz\" (UniqueName: \"kubernetes.io/projected/2cb23df5-8bb9-41fa-999e-6a563575b760-kube-api-access-lxnbz\") pod \"2cb23df5-8bb9-41fa-999e-6a563575b760\" (UID: \"2cb23df5-8bb9-41fa-999e-6a563575b760\") " Oct 02 10:17:03 crc kubenswrapper[4771]: I1002 10:17:03.703964 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb23df5-8bb9-41fa-999e-6a563575b760-ssh-key\") pod \"2cb23df5-8bb9-41fa-999e-6a563575b760\" (UID: \"2cb23df5-8bb9-41fa-999e-6a563575b760\") " Oct 02 10:17:03 crc kubenswrapper[4771]: I1002 10:17:03.711200 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cb23df5-8bb9-41fa-999e-6a563575b760-kube-api-access-lxnbz" (OuterVolumeSpecName: "kube-api-access-lxnbz") pod "2cb23df5-8bb9-41fa-999e-6a563575b760" (UID: "2cb23df5-8bb9-41fa-999e-6a563575b760"). InnerVolumeSpecName "kube-api-access-lxnbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:17:03 crc kubenswrapper[4771]: I1002 10:17:03.744374 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cb23df5-8bb9-41fa-999e-6a563575b760-inventory" (OuterVolumeSpecName: "inventory") pod "2cb23df5-8bb9-41fa-999e-6a563575b760" (UID: "2cb23df5-8bb9-41fa-999e-6a563575b760"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:17:03 crc kubenswrapper[4771]: I1002 10:17:03.748163 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cb23df5-8bb9-41fa-999e-6a563575b760-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2cb23df5-8bb9-41fa-999e-6a563575b760" (UID: "2cb23df5-8bb9-41fa-999e-6a563575b760"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:17:03 crc kubenswrapper[4771]: I1002 10:17:03.808574 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb23df5-8bb9-41fa-999e-6a563575b760-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:17:03 crc kubenswrapper[4771]: I1002 10:17:03.808839 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb23df5-8bb9-41fa-999e-6a563575b760-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:17:03 crc kubenswrapper[4771]: I1002 10:17:03.808854 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxnbz\" (UniqueName: \"kubernetes.io/projected/2cb23df5-8bb9-41fa-999e-6a563575b760-kube-api-access-lxnbz\") on node \"crc\" DevicePath \"\"" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.158390 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" event={"ID":"2cb23df5-8bb9-41fa-999e-6a563575b760","Type":"ContainerDied","Data":"cc3716856b8a945c624cf5b97e8ef2f101bf709dfe8ca891719d5cafd42bb76b"} Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.158427 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dsv2t" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.158431 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc3716856b8a945c624cf5b97e8ef2f101bf709dfe8ca891719d5cafd42bb76b" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.235274 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5"] Oct 02 10:17:04 crc kubenswrapper[4771]: E1002 10:17:04.235915 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cb23df5-8bb9-41fa-999e-6a563575b760" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.235939 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cb23df5-8bb9-41fa-999e-6a563575b760" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.236262 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cb23df5-8bb9-41fa-999e-6a563575b760" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.237383 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.240145 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.242494 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.242531 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.242527 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.249509 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5"] Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.277033 4771 scope.go:117] "RemoveContainer" containerID="90023531a6d5319f53f2bc581196dbbda23b992e70a511f264d7fec38a75689a" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.309768 4771 scope.go:117] "RemoveContainer" containerID="ca28488797bd03e21da30abbb0b1c3cc7af3a0ee4347241a01802f98eb3a6ffc" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.323053 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b22f05d8-7526-4e04-bac0-cdd965cac0cf-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5\" (UID: \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.323169 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzgrf\" (UniqueName: \"kubernetes.io/projected/b22f05d8-7526-4e04-bac0-cdd965cac0cf-kube-api-access-jzgrf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5\" (UID: \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.323297 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b22f05d8-7526-4e04-bac0-cdd965cac0cf-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5\" (UID: \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.426188 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b22f05d8-7526-4e04-bac0-cdd965cac0cf-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5\" (UID: \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.426242 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzgrf\" (UniqueName: \"kubernetes.io/projected/b22f05d8-7526-4e04-bac0-cdd965cac0cf-kube-api-access-jzgrf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5\" (UID: \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.426325 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b22f05d8-7526-4e04-bac0-cdd965cac0cf-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5\" (UID: \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.430447 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b22f05d8-7526-4e04-bac0-cdd965cac0cf-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5\" (UID: \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.430520 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b22f05d8-7526-4e04-bac0-cdd965cac0cf-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5\" (UID: \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.442565 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzgrf\" (UniqueName: \"kubernetes.io/projected/b22f05d8-7526-4e04-bac0-cdd965cac0cf-kube-api-access-jzgrf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5\" (UID: \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" Oct 02 10:17:04 crc kubenswrapper[4771]: I1002 10:17:04.565451 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" Oct 02 10:17:05 crc kubenswrapper[4771]: I1002 10:17:05.095009 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5"] Oct 02 10:17:05 crc kubenswrapper[4771]: I1002 10:17:05.172302 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" event={"ID":"b22f05d8-7526-4e04-bac0-cdd965cac0cf","Type":"ContainerStarted","Data":"6584eb021a5c3fa1d4c382b83878bbbd0500553733abb952e344f964eefa641c"} Oct 02 10:17:06 crc kubenswrapper[4771]: I1002 10:17:06.184902 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" event={"ID":"b22f05d8-7526-4e04-bac0-cdd965cac0cf","Type":"ContainerStarted","Data":"37bfe7e371e7c56e66c7d2f32ddb6017c07924c288598ba094d5cb06343b6f80"} Oct 02 10:17:06 crc kubenswrapper[4771]: I1002 10:17:06.204400 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" podStartSLOduration=1.700207585 podStartE2EDuration="2.20438299s" podCreationTimestamp="2025-10-02 10:17:04 +0000 UTC" firstStartedPulling="2025-10-02 10:17:05.099808625 +0000 UTC m=+2412.747493692" lastFinishedPulling="2025-10-02 10:17:05.60398403 +0000 UTC m=+2413.251669097" observedRunningTime="2025-10-02 10:17:06.202187043 +0000 UTC m=+2413.849872110" watchObservedRunningTime="2025-10-02 10:17:06.20438299 +0000 UTC m=+2413.852068057" Oct 02 10:17:12 crc kubenswrapper[4771]: I1002 10:17:12.681165 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:17:12 crc kubenswrapper[4771]: E1002 10:17:12.682042 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:17:16 crc kubenswrapper[4771]: I1002 10:17:16.319117 4771 generic.go:334] "Generic (PLEG): container finished" podID="b22f05d8-7526-4e04-bac0-cdd965cac0cf" containerID="37bfe7e371e7c56e66c7d2f32ddb6017c07924c288598ba094d5cb06343b6f80" exitCode=0 Oct 02 10:17:16 crc kubenswrapper[4771]: I1002 10:17:16.319193 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" event={"ID":"b22f05d8-7526-4e04-bac0-cdd965cac0cf","Type":"ContainerDied","Data":"37bfe7e371e7c56e66c7d2f32ddb6017c07924c288598ba094d5cb06343b6f80"} Oct 02 10:17:17 crc kubenswrapper[4771]: I1002 10:17:17.919059 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" Oct 02 10:17:17 crc kubenswrapper[4771]: I1002 10:17:17.984913 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b22f05d8-7526-4e04-bac0-cdd965cac0cf-inventory\") pod \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\" (UID: \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\") " Oct 02 10:17:17 crc kubenswrapper[4771]: I1002 10:17:17.985162 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b22f05d8-7526-4e04-bac0-cdd965cac0cf-ssh-key\") pod \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\" (UID: \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\") " Oct 02 10:17:17 crc kubenswrapper[4771]: I1002 10:17:17.985282 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzgrf\" (UniqueName: \"kubernetes.io/projected/b22f05d8-7526-4e04-bac0-cdd965cac0cf-kube-api-access-jzgrf\") pod \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\" (UID: \"b22f05d8-7526-4e04-bac0-cdd965cac0cf\") " Oct 02 10:17:17 crc kubenswrapper[4771]: I1002 10:17:17.990755 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b22f05d8-7526-4e04-bac0-cdd965cac0cf-kube-api-access-jzgrf" (OuterVolumeSpecName: "kube-api-access-jzgrf") pod "b22f05d8-7526-4e04-bac0-cdd965cac0cf" (UID: "b22f05d8-7526-4e04-bac0-cdd965cac0cf"). InnerVolumeSpecName "kube-api-access-jzgrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.019673 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b22f05d8-7526-4e04-bac0-cdd965cac0cf-inventory" (OuterVolumeSpecName: "inventory") pod "b22f05d8-7526-4e04-bac0-cdd965cac0cf" (UID: "b22f05d8-7526-4e04-bac0-cdd965cac0cf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.032363 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b22f05d8-7526-4e04-bac0-cdd965cac0cf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b22f05d8-7526-4e04-bac0-cdd965cac0cf" (UID: "b22f05d8-7526-4e04-bac0-cdd965cac0cf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.088013 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b22f05d8-7526-4e04-bac0-cdd965cac0cf-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.088052 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzgrf\" (UniqueName: \"kubernetes.io/projected/b22f05d8-7526-4e04-bac0-cdd965cac0cf-kube-api-access-jzgrf\") on node \"crc\" DevicePath \"\"" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.088068 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b22f05d8-7526-4e04-bac0-cdd965cac0cf-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.342828 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" event={"ID":"b22f05d8-7526-4e04-bac0-cdd965cac0cf","Type":"ContainerDied","Data":"6584eb021a5c3fa1d4c382b83878bbbd0500553733abb952e344f964eefa641c"} Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.342889 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6584eb021a5c3fa1d4c382b83878bbbd0500553733abb952e344f964eefa641c" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.342953 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.446302 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79"] Oct 02 10:17:18 crc kubenswrapper[4771]: E1002 10:17:18.446750 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b22f05d8-7526-4e04-bac0-cdd965cac0cf" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.446769 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b22f05d8-7526-4e04-bac0-cdd965cac0cf" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.447066 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="b22f05d8-7526-4e04-bac0-cdd965cac0cf" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.447855 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.449725 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.450230 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.453859 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.453857 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.453933 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.453966 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.454059 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.454115 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.454162 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.466901 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79"] Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.497600 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.497656 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.497752 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.497891 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.497937 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.497971 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.497997 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.498046 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.498210 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.498262 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8bzs\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-kube-api-access-f8bzs\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.498332 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.498573 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.498601 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.498632 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.498664 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.498740 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.601441 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.601546 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.601576 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.601596 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.601636 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.601660 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.601730 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.601757 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8bzs\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-kube-api-access-f8bzs\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.601809 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.601906 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.601927 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.601957 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.601982 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.602027 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.602070 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.602104 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.608290 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.608553 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.608900 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.609466 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.609732 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.610307 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.610531 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.610530 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.611435 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.611621 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.612263 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.612516 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.612709 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.615415 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.616746 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.620984 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8bzs\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-kube-api-access-f8bzs\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9kh79\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:18 crc kubenswrapper[4771]: I1002 10:17:18.770542 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:17:19 crc kubenswrapper[4771]: I1002 10:17:19.295597 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79"] Oct 02 10:17:19 crc kubenswrapper[4771]: W1002 10:17:19.340418 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2cab357_e81c_45c9_b0dc_6bdf44dd6092.slice/crio-4e553ead6bb03c223a6a3f860f898c1456985cfe1fb1f0b56680589cdd526036 WatchSource:0}: Error finding container 4e553ead6bb03c223a6a3f860f898c1456985cfe1fb1f0b56680589cdd526036: Status 404 returned error can't find the container with id 4e553ead6bb03c223a6a3f860f898c1456985cfe1fb1f0b56680589cdd526036 Oct 02 10:17:19 crc kubenswrapper[4771]: I1002 10:17:19.370633 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" event={"ID":"b2cab357-e81c-45c9-b0dc-6bdf44dd6092","Type":"ContainerStarted","Data":"4e553ead6bb03c223a6a3f860f898c1456985cfe1fb1f0b56680589cdd526036"} Oct 02 10:17:20 crc kubenswrapper[4771]: I1002 10:17:20.382099 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" event={"ID":"b2cab357-e81c-45c9-b0dc-6bdf44dd6092","Type":"ContainerStarted","Data":"b7b03b2191d50be1a33c313050dce9c2d978a7445866e242a67dcb0789d6f16a"} Oct 02 10:17:20 crc kubenswrapper[4771]: I1002 10:17:20.410546 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" podStartSLOduration=1.945389188 podStartE2EDuration="2.410525277s" podCreationTimestamp="2025-10-02 10:17:18 +0000 UTC" firstStartedPulling="2025-10-02 10:17:19.345349607 +0000 UTC m=+2426.993034684" lastFinishedPulling="2025-10-02 10:17:19.810485706 +0000 UTC m=+2427.458170773" observedRunningTime="2025-10-02 10:17:20.400868885 +0000 UTC m=+2428.048553952" watchObservedRunningTime="2025-10-02 10:17:20.410525277 +0000 UTC m=+2428.058210344" Oct 02 10:17:25 crc kubenswrapper[4771]: I1002 10:17:25.683838 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:17:25 crc kubenswrapper[4771]: E1002 10:17:25.684661 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:17:40 crc kubenswrapper[4771]: I1002 10:17:40.681288 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:17:40 crc kubenswrapper[4771]: E1002 10:17:40.682457 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:17:53 crc kubenswrapper[4771]: I1002 10:17:53.702195 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:17:53 crc kubenswrapper[4771]: E1002 10:17:53.703036 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:18:07 crc kubenswrapper[4771]: I1002 10:18:07.682857 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:18:07 crc kubenswrapper[4771]: E1002 10:18:07.684758 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:18:11 crc kubenswrapper[4771]: I1002 10:18:11.932481 4771 generic.go:334] "Generic (PLEG): container finished" podID="b2cab357-e81c-45c9-b0dc-6bdf44dd6092" containerID="b7b03b2191d50be1a33c313050dce9c2d978a7445866e242a67dcb0789d6f16a" exitCode=0 Oct 02 10:18:11 crc kubenswrapper[4771]: I1002 10:18:11.932571 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" event={"ID":"b2cab357-e81c-45c9-b0dc-6bdf44dd6092","Type":"ContainerDied","Data":"b7b03b2191d50be1a33c313050dce9c2d978a7445866e242a67dcb0789d6f16a"} Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.444824 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455168 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8bzs\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-kube-api-access-f8bzs\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455212 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-nova-combined-ca-bundle\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455237 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-ssh-key\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455296 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455365 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-repo-setup-combined-ca-bundle\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455392 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455440 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-ovn-default-certs-0\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455478 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-telemetry-combined-ca-bundle\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455509 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-libvirt-combined-ca-bundle\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455539 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-ovn-combined-ca-bundle\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455594 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455627 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-bootstrap-combined-ca-bundle\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455646 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455724 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-neutron-metadata-combined-ca-bundle\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455790 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-telemetry-power-monitoring-combined-ca-bundle\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.455860 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-inventory\") pod \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\" (UID: \"b2cab357-e81c-45c9-b0dc-6bdf44dd6092\") " Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.468400 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.468467 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.468261 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-kube-api-access-f8bzs" (OuterVolumeSpecName: "kube-api-access-f8bzs") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "kube-api-access-f8bzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.469257 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.470305 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.472349 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.472434 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.473282 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.475514 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.475732 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.475773 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.479313 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.500305 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.508809 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.517769 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-inventory" (OuterVolumeSpecName: "inventory") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.519368 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b2cab357-e81c-45c9-b0dc-6bdf44dd6092" (UID: "b2cab357-e81c-45c9-b0dc-6bdf44dd6092"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.557930 4771 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.557960 4771 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.557975 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.557985 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8bzs\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-kube-api-access-f8bzs\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.557994 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.558002 4771 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.558010 4771 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.558020 4771 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.558031 4771 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.558042 4771 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.558051 4771 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.558059 4771 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.558068 4771 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.558079 4771 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.558094 4771 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.558103 4771 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b2cab357-e81c-45c9-b0dc-6bdf44dd6092-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.959142 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" event={"ID":"b2cab357-e81c-45c9-b0dc-6bdf44dd6092","Type":"ContainerDied","Data":"4e553ead6bb03c223a6a3f860f898c1456985cfe1fb1f0b56680589cdd526036"} Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.959509 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e553ead6bb03c223a6a3f860f898c1456985cfe1fb1f0b56680589cdd526036" Oct 02 10:18:13 crc kubenswrapper[4771]: I1002 10:18:13.959186 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9kh79" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.071425 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw"] Oct 02 10:18:14 crc kubenswrapper[4771]: E1002 10:18:14.071961 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2cab357-e81c-45c9-b0dc-6bdf44dd6092" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.071984 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2cab357-e81c-45c9-b0dc-6bdf44dd6092" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.072257 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2cab357-e81c-45c9-b0dc-6bdf44dd6092" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.073258 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.076786 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.078650 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.078721 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.079438 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.080411 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.092316 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw"] Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.274492 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.274599 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txj5c\" (UniqueName: \"kubernetes.io/projected/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-kube-api-access-txj5c\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.274649 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.275783 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.275830 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.378450 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txj5c\" (UniqueName: \"kubernetes.io/projected/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-kube-api-access-txj5c\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.378533 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.378631 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.378666 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.378720 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.379683 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.385088 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.386613 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.388667 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.396885 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txj5c\" (UniqueName: \"kubernetes.io/projected/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-kube-api-access-txj5c\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ctttw\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:14 crc kubenswrapper[4771]: I1002 10:18:14.694169 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:18:15 crc kubenswrapper[4771]: I1002 10:18:15.238027 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw"] Oct 02 10:18:15 crc kubenswrapper[4771]: I1002 10:18:15.989757 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" event={"ID":"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d","Type":"ContainerStarted","Data":"04377d42ee7acb75cfe2c767a89c0c8de37c695e14fc9a8f5917f6dcd995cb74"} Oct 02 10:18:17 crc kubenswrapper[4771]: I1002 10:18:17.001555 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" event={"ID":"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d","Type":"ContainerStarted","Data":"018d457ec91d41f0a3ad51135f80f6a641d29e6d1967a4dce9ecfe08a6a46a03"} Oct 02 10:18:22 crc kubenswrapper[4771]: I1002 10:18:22.681587 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:18:22 crc kubenswrapper[4771]: E1002 10:18:22.682678 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:18:37 crc kubenswrapper[4771]: I1002 10:18:37.681643 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:18:37 crc kubenswrapper[4771]: E1002 10:18:37.682599 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:18:48 crc kubenswrapper[4771]: I1002 10:18:48.681492 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:18:48 crc kubenswrapper[4771]: E1002 10:18:48.682616 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:19:00 crc kubenswrapper[4771]: I1002 10:19:00.681922 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:19:00 crc kubenswrapper[4771]: E1002 10:19:00.683356 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:19:14 crc kubenswrapper[4771]: I1002 10:19:14.681426 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:19:14 crc kubenswrapper[4771]: E1002 10:19:14.682147 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:19:27 crc kubenswrapper[4771]: I1002 10:19:27.844807 4771 generic.go:334] "Generic (PLEG): container finished" podID="fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d" containerID="018d457ec91d41f0a3ad51135f80f6a641d29e6d1967a4dce9ecfe08a6a46a03" exitCode=0 Oct 02 10:19:27 crc kubenswrapper[4771]: I1002 10:19:27.844921 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" event={"ID":"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d","Type":"ContainerDied","Data":"018d457ec91d41f0a3ad51135f80f6a641d29e6d1967a4dce9ecfe08a6a46a03"} Oct 02 10:19:28 crc kubenswrapper[4771]: I1002 10:19:28.682730 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:19:28 crc kubenswrapper[4771]: E1002 10:19:28.683250 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.458069 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.592443 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-inventory\") pod \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.592601 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txj5c\" (UniqueName: \"kubernetes.io/projected/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-kube-api-access-txj5c\") pod \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.592839 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ovn-combined-ca-bundle\") pod \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.592871 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ovncontroller-config-0\") pod \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.592960 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ssh-key\") pod \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\" (UID: \"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d\") " Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.597933 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d" (UID: "fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.598311 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-kube-api-access-txj5c" (OuterVolumeSpecName: "kube-api-access-txj5c") pod "fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d" (UID: "fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d"). InnerVolumeSpecName "kube-api-access-txj5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.631538 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d" (UID: "fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.637791 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d" (UID: "fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.638222 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-inventory" (OuterVolumeSpecName: "inventory") pod "fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d" (UID: "fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.696427 4771 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.696712 4771 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.696725 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.696735 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.696744 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txj5c\" (UniqueName: \"kubernetes.io/projected/fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d-kube-api-access-txj5c\") on node \"crc\" DevicePath \"\"" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.867544 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" event={"ID":"fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d","Type":"ContainerDied","Data":"04377d42ee7acb75cfe2c767a89c0c8de37c695e14fc9a8f5917f6dcd995cb74"} Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.867821 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04377d42ee7acb75cfe2c767a89c0c8de37c695e14fc9a8f5917f6dcd995cb74" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.867661 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ctttw" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.969564 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg"] Oct 02 10:19:29 crc kubenswrapper[4771]: E1002 10:19:29.970147 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.970168 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.970457 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.971294 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.974415 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.974755 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.975094 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.975254 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.975307 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.980199 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 02 10:19:29 crc kubenswrapper[4771]: I1002 10:19:29.980923 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg"] Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.122806 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.123164 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.123420 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.123766 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwdm6\" (UniqueName: \"kubernetes.io/projected/6697b08a-190c-48d2-9528-4cccd4efb520-kube-api-access-wwdm6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.123936 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.124066 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.226506 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwdm6\" (UniqueName: \"kubernetes.io/projected/6697b08a-190c-48d2-9528-4cccd4efb520-kube-api-access-wwdm6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.226562 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.226597 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.226659 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.226759 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.226847 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.231352 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.231620 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.232591 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.233573 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.240716 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.254879 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwdm6\" (UniqueName: \"kubernetes.io/projected/6697b08a-190c-48d2-9528-4cccd4efb520-kube-api-access-wwdm6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.295283 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:19:30 crc kubenswrapper[4771]: I1002 10:19:30.880060 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg"] Oct 02 10:19:31 crc kubenswrapper[4771]: I1002 10:19:31.894750 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" event={"ID":"6697b08a-190c-48d2-9528-4cccd4efb520","Type":"ContainerStarted","Data":"95b951afca5ef66b0186212b3f8f97896b88157520b3352a9c80230d7d4e4571"} Oct 02 10:19:31 crc kubenswrapper[4771]: I1002 10:19:31.895108 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" event={"ID":"6697b08a-190c-48d2-9528-4cccd4efb520","Type":"ContainerStarted","Data":"a229a63013f53531de57733edb692601e54a321af263797042e677733a461da3"} Oct 02 10:19:31 crc kubenswrapper[4771]: I1002 10:19:31.915262 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" podStartSLOduration=2.481310711 podStartE2EDuration="2.915238548s" podCreationTimestamp="2025-10-02 10:19:29 +0000 UTC" firstStartedPulling="2025-10-02 10:19:30.884413422 +0000 UTC m=+2558.532098499" lastFinishedPulling="2025-10-02 10:19:31.318341249 +0000 UTC m=+2558.966026336" observedRunningTime="2025-10-02 10:19:31.913749649 +0000 UTC m=+2559.561434716" watchObservedRunningTime="2025-10-02 10:19:31.915238548 +0000 UTC m=+2559.562923615" Oct 02 10:19:39 crc kubenswrapper[4771]: I1002 10:19:39.684309 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:19:39 crc kubenswrapper[4771]: E1002 10:19:39.685410 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:19:51 crc kubenswrapper[4771]: I1002 10:19:51.681298 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:19:52 crc kubenswrapper[4771]: I1002 10:19:52.133655 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"f79d932eb1d7bfadfdeb97c1d67c7b20b8fc5e7820bfc2dce0c429b41092abd6"} Oct 02 10:20:26 crc kubenswrapper[4771]: I1002 10:20:26.539368 4771 generic.go:334] "Generic (PLEG): container finished" podID="6697b08a-190c-48d2-9528-4cccd4efb520" containerID="95b951afca5ef66b0186212b3f8f97896b88157520b3352a9c80230d7d4e4571" exitCode=0 Oct 02 10:20:26 crc kubenswrapper[4771]: I1002 10:20:26.539551 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" event={"ID":"6697b08a-190c-48d2-9528-4cccd4efb520","Type":"ContainerDied","Data":"95b951afca5ef66b0186212b3f8f97896b88157520b3352a9c80230d7d4e4571"} Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.044720 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.189857 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-inventory\") pod \"6697b08a-190c-48d2-9528-4cccd4efb520\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.189990 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-ssh-key\") pod \"6697b08a-190c-48d2-9528-4cccd4efb520\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.190930 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-neutron-ovn-metadata-agent-neutron-config-0\") pod \"6697b08a-190c-48d2-9528-4cccd4efb520\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.191051 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-neutron-metadata-combined-ca-bundle\") pod \"6697b08a-190c-48d2-9528-4cccd4efb520\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.191077 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-nova-metadata-neutron-config-0\") pod \"6697b08a-190c-48d2-9528-4cccd4efb520\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.191100 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwdm6\" (UniqueName: \"kubernetes.io/projected/6697b08a-190c-48d2-9528-4cccd4efb520-kube-api-access-wwdm6\") pod \"6697b08a-190c-48d2-9528-4cccd4efb520\" (UID: \"6697b08a-190c-48d2-9528-4cccd4efb520\") " Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.195398 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "6697b08a-190c-48d2-9528-4cccd4efb520" (UID: "6697b08a-190c-48d2-9528-4cccd4efb520"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.219199 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6697b08a-190c-48d2-9528-4cccd4efb520-kube-api-access-wwdm6" (OuterVolumeSpecName: "kube-api-access-wwdm6") pod "6697b08a-190c-48d2-9528-4cccd4efb520" (UID: "6697b08a-190c-48d2-9528-4cccd4efb520"). InnerVolumeSpecName "kube-api-access-wwdm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.223468 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "6697b08a-190c-48d2-9528-4cccd4efb520" (UID: "6697b08a-190c-48d2-9528-4cccd4efb520"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.228566 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "6697b08a-190c-48d2-9528-4cccd4efb520" (UID: "6697b08a-190c-48d2-9528-4cccd4efb520"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.232860 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6697b08a-190c-48d2-9528-4cccd4efb520" (UID: "6697b08a-190c-48d2-9528-4cccd4efb520"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.234788 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-inventory" (OuterVolumeSpecName: "inventory") pod "6697b08a-190c-48d2-9528-4cccd4efb520" (UID: "6697b08a-190c-48d2-9528-4cccd4efb520"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.293779 4771 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.293810 4771 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.293820 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwdm6\" (UniqueName: \"kubernetes.io/projected/6697b08a-190c-48d2-9528-4cccd4efb520-kube-api-access-wwdm6\") on node \"crc\" DevicePath \"\"" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.293831 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.293839 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.293847 4771 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6697b08a-190c-48d2-9528-4cccd4efb520-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.559539 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" event={"ID":"6697b08a-190c-48d2-9528-4cccd4efb520","Type":"ContainerDied","Data":"a229a63013f53531de57733edb692601e54a321af263797042e677733a461da3"} Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.559852 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a229a63013f53531de57733edb692601e54a321af263797042e677733a461da3" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.559643 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.712630 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x"] Oct 02 10:20:28 crc kubenswrapper[4771]: E1002 10:20:28.713388 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6697b08a-190c-48d2-9528-4cccd4efb520" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.713409 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6697b08a-190c-48d2-9528-4cccd4efb520" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.713682 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="6697b08a-190c-48d2-9528-4cccd4efb520" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.714706 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.720543 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.720612 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.720690 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.720755 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.720550 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.773654 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x"] Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.815261 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.815659 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.815822 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.816057 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg29v\" (UniqueName: \"kubernetes.io/projected/7c7c8501-8ae8-47d5-b947-8b9ac02be829-kube-api-access-qg29v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.816143 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.918390 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg29v\" (UniqueName: \"kubernetes.io/projected/7c7c8501-8ae8-47d5-b947-8b9ac02be829-kube-api-access-qg29v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.918434 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.918498 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.918571 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.918612 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.924164 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.924815 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.924960 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.926700 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:28 crc kubenswrapper[4771]: I1002 10:20:28.937353 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg29v\" (UniqueName: \"kubernetes.io/projected/7c7c8501-8ae8-47d5-b947-8b9ac02be829-kube-api-access-qg29v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:29 crc kubenswrapper[4771]: I1002 10:20:29.084654 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:20:29 crc kubenswrapper[4771]: I1002 10:20:29.672900 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x"] Oct 02 10:20:30 crc kubenswrapper[4771]: I1002 10:20:30.583712 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" event={"ID":"7c7c8501-8ae8-47d5-b947-8b9ac02be829","Type":"ContainerStarted","Data":"cff854af87673a42acffe312f14a7da6bfd71b67e5033c1f7ed6aaf240008239"} Oct 02 10:20:31 crc kubenswrapper[4771]: I1002 10:20:31.597662 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" event={"ID":"7c7c8501-8ae8-47d5-b947-8b9ac02be829","Type":"ContainerStarted","Data":"e85ee3edcadfe21735bfa9f46a426144b26dc1688dfee8a8b9c3b22da3a948cf"} Oct 02 10:20:31 crc kubenswrapper[4771]: I1002 10:20:31.626383 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" podStartSLOduration=3.016532683 podStartE2EDuration="3.6263549s" podCreationTimestamp="2025-10-02 10:20:28 +0000 UTC" firstStartedPulling="2025-10-02 10:20:29.669762969 +0000 UTC m=+2617.317448066" lastFinishedPulling="2025-10-02 10:20:30.279585216 +0000 UTC m=+2617.927270283" observedRunningTime="2025-10-02 10:20:31.613185956 +0000 UTC m=+2619.260871023" watchObservedRunningTime="2025-10-02 10:20:31.6263549 +0000 UTC m=+2619.274039997" Oct 02 10:22:12 crc kubenswrapper[4771]: I1002 10:22:12.146120 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:22:12 crc kubenswrapper[4771]: I1002 10:22:12.146674 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:22:14 crc kubenswrapper[4771]: I1002 10:22:14.809904 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-brq9r"] Oct 02 10:22:14 crc kubenswrapper[4771]: I1002 10:22:14.814790 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:14 crc kubenswrapper[4771]: I1002 10:22:14.830329 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-brq9r"] Oct 02 10:22:14 crc kubenswrapper[4771]: I1002 10:22:14.934246 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c72f989c-3f0c-4f2d-9344-2439d41f99db-catalog-content\") pod \"redhat-operators-brq9r\" (UID: \"c72f989c-3f0c-4f2d-9344-2439d41f99db\") " pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:14 crc kubenswrapper[4771]: I1002 10:22:14.934704 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c72f989c-3f0c-4f2d-9344-2439d41f99db-utilities\") pod \"redhat-operators-brq9r\" (UID: \"c72f989c-3f0c-4f2d-9344-2439d41f99db\") " pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:14 crc kubenswrapper[4771]: I1002 10:22:14.934848 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5xcb\" (UniqueName: \"kubernetes.io/projected/c72f989c-3f0c-4f2d-9344-2439d41f99db-kube-api-access-j5xcb\") pod \"redhat-operators-brq9r\" (UID: \"c72f989c-3f0c-4f2d-9344-2439d41f99db\") " pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:15 crc kubenswrapper[4771]: I1002 10:22:15.036343 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c72f989c-3f0c-4f2d-9344-2439d41f99db-catalog-content\") pod \"redhat-operators-brq9r\" (UID: \"c72f989c-3f0c-4f2d-9344-2439d41f99db\") " pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:15 crc kubenswrapper[4771]: I1002 10:22:15.036449 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c72f989c-3f0c-4f2d-9344-2439d41f99db-utilities\") pod \"redhat-operators-brq9r\" (UID: \"c72f989c-3f0c-4f2d-9344-2439d41f99db\") " pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:15 crc kubenswrapper[4771]: I1002 10:22:15.036516 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5xcb\" (UniqueName: \"kubernetes.io/projected/c72f989c-3f0c-4f2d-9344-2439d41f99db-kube-api-access-j5xcb\") pod \"redhat-operators-brq9r\" (UID: \"c72f989c-3f0c-4f2d-9344-2439d41f99db\") " pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:15 crc kubenswrapper[4771]: I1002 10:22:15.036820 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c72f989c-3f0c-4f2d-9344-2439d41f99db-catalog-content\") pod \"redhat-operators-brq9r\" (UID: \"c72f989c-3f0c-4f2d-9344-2439d41f99db\") " pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:15 crc kubenswrapper[4771]: I1002 10:22:15.037033 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c72f989c-3f0c-4f2d-9344-2439d41f99db-utilities\") pod \"redhat-operators-brq9r\" (UID: \"c72f989c-3f0c-4f2d-9344-2439d41f99db\") " pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:15 crc kubenswrapper[4771]: I1002 10:22:15.056491 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5xcb\" (UniqueName: \"kubernetes.io/projected/c72f989c-3f0c-4f2d-9344-2439d41f99db-kube-api-access-j5xcb\") pod \"redhat-operators-brq9r\" (UID: \"c72f989c-3f0c-4f2d-9344-2439d41f99db\") " pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:15 crc kubenswrapper[4771]: I1002 10:22:15.160526 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:15 crc kubenswrapper[4771]: I1002 10:22:15.680718 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-brq9r"] Oct 02 10:22:15 crc kubenswrapper[4771]: I1002 10:22:15.900549 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brq9r" event={"ID":"c72f989c-3f0c-4f2d-9344-2439d41f99db","Type":"ContainerStarted","Data":"ea0ebf522688d5706d256e2bbde961248c893703b734133ddd7b35be9cbaf5ba"} Oct 02 10:22:16 crc kubenswrapper[4771]: I1002 10:22:16.923766 4771 generic.go:334] "Generic (PLEG): container finished" podID="c72f989c-3f0c-4f2d-9344-2439d41f99db" containerID="6efb166697370d12d1fe784d1efe64aef8b346ffbca25a7bb0befa185c425046" exitCode=0 Oct 02 10:22:16 crc kubenswrapper[4771]: I1002 10:22:16.923824 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brq9r" event={"ID":"c72f989c-3f0c-4f2d-9344-2439d41f99db","Type":"ContainerDied","Data":"6efb166697370d12d1fe784d1efe64aef8b346ffbca25a7bb0befa185c425046"} Oct 02 10:22:16 crc kubenswrapper[4771]: I1002 10:22:16.928438 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:22:18 crc kubenswrapper[4771]: I1002 10:22:18.957163 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brq9r" event={"ID":"c72f989c-3f0c-4f2d-9344-2439d41f99db","Type":"ContainerStarted","Data":"5a18f06d57b48df6579bf8cba6f171cdab20a2d05464c3f3dacd33567355b18f"} Oct 02 10:22:22 crc kubenswrapper[4771]: I1002 10:22:22.003221 4771 generic.go:334] "Generic (PLEG): container finished" podID="c72f989c-3f0c-4f2d-9344-2439d41f99db" containerID="5a18f06d57b48df6579bf8cba6f171cdab20a2d05464c3f3dacd33567355b18f" exitCode=0 Oct 02 10:22:22 crc kubenswrapper[4771]: I1002 10:22:22.003685 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brq9r" event={"ID":"c72f989c-3f0c-4f2d-9344-2439d41f99db","Type":"ContainerDied","Data":"5a18f06d57b48df6579bf8cba6f171cdab20a2d05464c3f3dacd33567355b18f"} Oct 02 10:22:23 crc kubenswrapper[4771]: I1002 10:22:23.024176 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brq9r" event={"ID":"c72f989c-3f0c-4f2d-9344-2439d41f99db","Type":"ContainerStarted","Data":"95c82167e5e6073560836f2db9b42f946b1d0430137a5db1bd110836b3842eec"} Oct 02 10:22:23 crc kubenswrapper[4771]: I1002 10:22:23.064509 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-brq9r" podStartSLOduration=3.573936326 podStartE2EDuration="9.06447767s" podCreationTimestamp="2025-10-02 10:22:14 +0000 UTC" firstStartedPulling="2025-10-02 10:22:16.928231426 +0000 UTC m=+2724.575916493" lastFinishedPulling="2025-10-02 10:22:22.41877277 +0000 UTC m=+2730.066457837" observedRunningTime="2025-10-02 10:22:23.048315089 +0000 UTC m=+2730.696000156" watchObservedRunningTime="2025-10-02 10:22:23.06447767 +0000 UTC m=+2730.712162737" Oct 02 10:22:25 crc kubenswrapper[4771]: I1002 10:22:25.160963 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:25 crc kubenswrapper[4771]: I1002 10:22:25.161315 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:26 crc kubenswrapper[4771]: I1002 10:22:26.223982 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-brq9r" podUID="c72f989c-3f0c-4f2d-9344-2439d41f99db" containerName="registry-server" probeResult="failure" output=< Oct 02 10:22:26 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:22:26 crc kubenswrapper[4771]: > Oct 02 10:22:36 crc kubenswrapper[4771]: I1002 10:22:36.214874 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-brq9r" podUID="c72f989c-3f0c-4f2d-9344-2439d41f99db" containerName="registry-server" probeResult="failure" output=< Oct 02 10:22:36 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:22:36 crc kubenswrapper[4771]: > Oct 02 10:22:42 crc kubenswrapper[4771]: I1002 10:22:42.146247 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:22:42 crc kubenswrapper[4771]: I1002 10:22:42.147211 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:22:45 crc kubenswrapper[4771]: I1002 10:22:45.213416 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:45 crc kubenswrapper[4771]: I1002 10:22:45.266928 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:46 crc kubenswrapper[4771]: I1002 10:22:46.005296 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-brq9r"] Oct 02 10:22:46 crc kubenswrapper[4771]: I1002 10:22:46.312009 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-brq9r" podUID="c72f989c-3f0c-4f2d-9344-2439d41f99db" containerName="registry-server" containerID="cri-o://95c82167e5e6073560836f2db9b42f946b1d0430137a5db1bd110836b3842eec" gracePeriod=2 Oct 02 10:22:46 crc kubenswrapper[4771]: I1002 10:22:46.839272 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:46 crc kubenswrapper[4771]: I1002 10:22:46.871786 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5xcb\" (UniqueName: \"kubernetes.io/projected/c72f989c-3f0c-4f2d-9344-2439d41f99db-kube-api-access-j5xcb\") pod \"c72f989c-3f0c-4f2d-9344-2439d41f99db\" (UID: \"c72f989c-3f0c-4f2d-9344-2439d41f99db\") " Oct 02 10:22:46 crc kubenswrapper[4771]: I1002 10:22:46.871899 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c72f989c-3f0c-4f2d-9344-2439d41f99db-catalog-content\") pod \"c72f989c-3f0c-4f2d-9344-2439d41f99db\" (UID: \"c72f989c-3f0c-4f2d-9344-2439d41f99db\") " Oct 02 10:22:46 crc kubenswrapper[4771]: I1002 10:22:46.872037 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c72f989c-3f0c-4f2d-9344-2439d41f99db-utilities\") pod \"c72f989c-3f0c-4f2d-9344-2439d41f99db\" (UID: \"c72f989c-3f0c-4f2d-9344-2439d41f99db\") " Oct 02 10:22:46 crc kubenswrapper[4771]: I1002 10:22:46.872654 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c72f989c-3f0c-4f2d-9344-2439d41f99db-utilities" (OuterVolumeSpecName: "utilities") pod "c72f989c-3f0c-4f2d-9344-2439d41f99db" (UID: "c72f989c-3f0c-4f2d-9344-2439d41f99db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:22:46 crc kubenswrapper[4771]: I1002 10:22:46.873002 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c72f989c-3f0c-4f2d-9344-2439d41f99db-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:22:46 crc kubenswrapper[4771]: I1002 10:22:46.902522 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c72f989c-3f0c-4f2d-9344-2439d41f99db-kube-api-access-j5xcb" (OuterVolumeSpecName: "kube-api-access-j5xcb") pod "c72f989c-3f0c-4f2d-9344-2439d41f99db" (UID: "c72f989c-3f0c-4f2d-9344-2439d41f99db"). InnerVolumeSpecName "kube-api-access-j5xcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:22:46 crc kubenswrapper[4771]: I1002 10:22:46.974479 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c72f989c-3f0c-4f2d-9344-2439d41f99db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c72f989c-3f0c-4f2d-9344-2439d41f99db" (UID: "c72f989c-3f0c-4f2d-9344-2439d41f99db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:22:46 crc kubenswrapper[4771]: I1002 10:22:46.974788 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5xcb\" (UniqueName: \"kubernetes.io/projected/c72f989c-3f0c-4f2d-9344-2439d41f99db-kube-api-access-j5xcb\") on node \"crc\" DevicePath \"\"" Oct 02 10:22:46 crc kubenswrapper[4771]: I1002 10:22:46.974804 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c72f989c-3f0c-4f2d-9344-2439d41f99db-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.323443 4771 generic.go:334] "Generic (PLEG): container finished" podID="c72f989c-3f0c-4f2d-9344-2439d41f99db" containerID="95c82167e5e6073560836f2db9b42f946b1d0430137a5db1bd110836b3842eec" exitCode=0 Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.323504 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brq9r" event={"ID":"c72f989c-3f0c-4f2d-9344-2439d41f99db","Type":"ContainerDied","Data":"95c82167e5e6073560836f2db9b42f946b1d0430137a5db1bd110836b3842eec"} Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.323542 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brq9r" event={"ID":"c72f989c-3f0c-4f2d-9344-2439d41f99db","Type":"ContainerDied","Data":"ea0ebf522688d5706d256e2bbde961248c893703b734133ddd7b35be9cbaf5ba"} Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.323553 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brq9r" Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.323562 4771 scope.go:117] "RemoveContainer" containerID="95c82167e5e6073560836f2db9b42f946b1d0430137a5db1bd110836b3842eec" Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.347910 4771 scope.go:117] "RemoveContainer" containerID="5a18f06d57b48df6579bf8cba6f171cdab20a2d05464c3f3dacd33567355b18f" Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.361779 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-brq9r"] Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.369244 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-brq9r"] Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.382784 4771 scope.go:117] "RemoveContainer" containerID="6efb166697370d12d1fe784d1efe64aef8b346ffbca25a7bb0befa185c425046" Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.442193 4771 scope.go:117] "RemoveContainer" containerID="95c82167e5e6073560836f2db9b42f946b1d0430137a5db1bd110836b3842eec" Oct 02 10:22:47 crc kubenswrapper[4771]: E1002 10:22:47.442874 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95c82167e5e6073560836f2db9b42f946b1d0430137a5db1bd110836b3842eec\": container with ID starting with 95c82167e5e6073560836f2db9b42f946b1d0430137a5db1bd110836b3842eec not found: ID does not exist" containerID="95c82167e5e6073560836f2db9b42f946b1d0430137a5db1bd110836b3842eec" Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.442977 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95c82167e5e6073560836f2db9b42f946b1d0430137a5db1bd110836b3842eec"} err="failed to get container status \"95c82167e5e6073560836f2db9b42f946b1d0430137a5db1bd110836b3842eec\": rpc error: code = NotFound desc = could not find container \"95c82167e5e6073560836f2db9b42f946b1d0430137a5db1bd110836b3842eec\": container with ID starting with 95c82167e5e6073560836f2db9b42f946b1d0430137a5db1bd110836b3842eec not found: ID does not exist" Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.443005 4771 scope.go:117] "RemoveContainer" containerID="5a18f06d57b48df6579bf8cba6f171cdab20a2d05464c3f3dacd33567355b18f" Oct 02 10:22:47 crc kubenswrapper[4771]: E1002 10:22:47.443243 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a18f06d57b48df6579bf8cba6f171cdab20a2d05464c3f3dacd33567355b18f\": container with ID starting with 5a18f06d57b48df6579bf8cba6f171cdab20a2d05464c3f3dacd33567355b18f not found: ID does not exist" containerID="5a18f06d57b48df6579bf8cba6f171cdab20a2d05464c3f3dacd33567355b18f" Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.443368 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a18f06d57b48df6579bf8cba6f171cdab20a2d05464c3f3dacd33567355b18f"} err="failed to get container status \"5a18f06d57b48df6579bf8cba6f171cdab20a2d05464c3f3dacd33567355b18f\": rpc error: code = NotFound desc = could not find container \"5a18f06d57b48df6579bf8cba6f171cdab20a2d05464c3f3dacd33567355b18f\": container with ID starting with 5a18f06d57b48df6579bf8cba6f171cdab20a2d05464c3f3dacd33567355b18f not found: ID does not exist" Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.443405 4771 scope.go:117] "RemoveContainer" containerID="6efb166697370d12d1fe784d1efe64aef8b346ffbca25a7bb0befa185c425046" Oct 02 10:22:47 crc kubenswrapper[4771]: E1002 10:22:47.443862 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6efb166697370d12d1fe784d1efe64aef8b346ffbca25a7bb0befa185c425046\": container with ID starting with 6efb166697370d12d1fe784d1efe64aef8b346ffbca25a7bb0befa185c425046 not found: ID does not exist" containerID="6efb166697370d12d1fe784d1efe64aef8b346ffbca25a7bb0befa185c425046" Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.443907 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6efb166697370d12d1fe784d1efe64aef8b346ffbca25a7bb0befa185c425046"} err="failed to get container status \"6efb166697370d12d1fe784d1efe64aef8b346ffbca25a7bb0befa185c425046\": rpc error: code = NotFound desc = could not find container \"6efb166697370d12d1fe784d1efe64aef8b346ffbca25a7bb0befa185c425046\": container with ID starting with 6efb166697370d12d1fe784d1efe64aef8b346ffbca25a7bb0befa185c425046 not found: ID does not exist" Oct 02 10:22:47 crc kubenswrapper[4771]: I1002 10:22:47.696598 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c72f989c-3f0c-4f2d-9344-2439d41f99db" path="/var/lib/kubelet/pods/c72f989c-3f0c-4f2d-9344-2439d41f99db/volumes" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.220566 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-27nfg"] Oct 02 10:22:48 crc kubenswrapper[4771]: E1002 10:22:48.221089 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c72f989c-3f0c-4f2d-9344-2439d41f99db" containerName="extract-content" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.221105 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="c72f989c-3f0c-4f2d-9344-2439d41f99db" containerName="extract-content" Oct 02 10:22:48 crc kubenswrapper[4771]: E1002 10:22:48.221151 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c72f989c-3f0c-4f2d-9344-2439d41f99db" containerName="extract-utilities" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.221161 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="c72f989c-3f0c-4f2d-9344-2439d41f99db" containerName="extract-utilities" Oct 02 10:22:48 crc kubenswrapper[4771]: E1002 10:22:48.221191 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c72f989c-3f0c-4f2d-9344-2439d41f99db" containerName="registry-server" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.221198 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="c72f989c-3f0c-4f2d-9344-2439d41f99db" containerName="registry-server" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.221441 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="c72f989c-3f0c-4f2d-9344-2439d41f99db" containerName="registry-server" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.224045 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.237528 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-27nfg"] Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.305755 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4trh\" (UniqueName: \"kubernetes.io/projected/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-kube-api-access-q4trh\") pod \"certified-operators-27nfg\" (UID: \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\") " pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.305818 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-utilities\") pod \"certified-operators-27nfg\" (UID: \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\") " pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.305861 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-catalog-content\") pod \"certified-operators-27nfg\" (UID: \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\") " pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.409395 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4trh\" (UniqueName: \"kubernetes.io/projected/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-kube-api-access-q4trh\") pod \"certified-operators-27nfg\" (UID: \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\") " pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.409456 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-utilities\") pod \"certified-operators-27nfg\" (UID: \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\") " pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.409499 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-catalog-content\") pod \"certified-operators-27nfg\" (UID: \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\") " pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.409980 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-catalog-content\") pod \"certified-operators-27nfg\" (UID: \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\") " pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.410096 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-utilities\") pod \"certified-operators-27nfg\" (UID: \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\") " pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.457993 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4trh\" (UniqueName: \"kubernetes.io/projected/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-kube-api-access-q4trh\") pod \"certified-operators-27nfg\" (UID: \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\") " pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:48 crc kubenswrapper[4771]: I1002 10:22:48.555766 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:49 crc kubenswrapper[4771]: I1002 10:22:49.069251 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-27nfg"] Oct 02 10:22:49 crc kubenswrapper[4771]: I1002 10:22:49.349325 4771 generic.go:334] "Generic (PLEG): container finished" podID="54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" containerID="442bc533923f383c3b002785f20305583effce25fe70db590dc541469972a504" exitCode=0 Oct 02 10:22:49 crc kubenswrapper[4771]: I1002 10:22:49.349378 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27nfg" event={"ID":"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5","Type":"ContainerDied","Data":"442bc533923f383c3b002785f20305583effce25fe70db590dc541469972a504"} Oct 02 10:22:49 crc kubenswrapper[4771]: I1002 10:22:49.349413 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27nfg" event={"ID":"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5","Type":"ContainerStarted","Data":"3097f68bf074122ea7f47ddc5d8f37cac2d974428e2a8744d33c8d3c2426a4ed"} Oct 02 10:22:50 crc kubenswrapper[4771]: I1002 10:22:50.364582 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27nfg" event={"ID":"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5","Type":"ContainerStarted","Data":"561267b5b6c96e97e5b4e40da6e0637be09487241a184febd3e8805578825543"} Oct 02 10:22:51 crc kubenswrapper[4771]: I1002 10:22:51.375448 4771 generic.go:334] "Generic (PLEG): container finished" podID="54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" containerID="561267b5b6c96e97e5b4e40da6e0637be09487241a184febd3e8805578825543" exitCode=0 Oct 02 10:22:51 crc kubenswrapper[4771]: I1002 10:22:51.375529 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27nfg" event={"ID":"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5","Type":"ContainerDied","Data":"561267b5b6c96e97e5b4e40da6e0637be09487241a184febd3e8805578825543"} Oct 02 10:22:52 crc kubenswrapper[4771]: I1002 10:22:52.393335 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27nfg" event={"ID":"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5","Type":"ContainerStarted","Data":"c39f8ac7793311055928435ec5d285460f5c9573b24034da38d8170466d6d911"} Oct 02 10:22:52 crc kubenswrapper[4771]: I1002 10:22:52.415627 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-27nfg" podStartSLOduration=1.9122577139999999 podStartE2EDuration="4.415611342s" podCreationTimestamp="2025-10-02 10:22:48 +0000 UTC" firstStartedPulling="2025-10-02 10:22:49.35130602 +0000 UTC m=+2756.998991097" lastFinishedPulling="2025-10-02 10:22:51.854659668 +0000 UTC m=+2759.502344725" observedRunningTime="2025-10-02 10:22:52.413083073 +0000 UTC m=+2760.060768140" watchObservedRunningTime="2025-10-02 10:22:52.415611342 +0000 UTC m=+2760.063296409" Oct 02 10:22:58 crc kubenswrapper[4771]: I1002 10:22:58.556163 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:58 crc kubenswrapper[4771]: I1002 10:22:58.556668 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:58 crc kubenswrapper[4771]: I1002 10:22:58.621324 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:59 crc kubenswrapper[4771]: I1002 10:22:59.516188 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:22:59 crc kubenswrapper[4771]: I1002 10:22:59.570216 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-27nfg"] Oct 02 10:23:01 crc kubenswrapper[4771]: I1002 10:23:01.486438 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-27nfg" podUID="54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" containerName="registry-server" containerID="cri-o://c39f8ac7793311055928435ec5d285460f5c9573b24034da38d8170466d6d911" gracePeriod=2 Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.074744 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.261512 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-utilities\") pod \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\" (UID: \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\") " Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.261573 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-catalog-content\") pod \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\" (UID: \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\") " Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.261735 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4trh\" (UniqueName: \"kubernetes.io/projected/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-kube-api-access-q4trh\") pod \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\" (UID: \"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5\") " Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.263258 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-utilities" (OuterVolumeSpecName: "utilities") pod "54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" (UID: "54ed9bd6-44dd-42fc-8640-8e9bd2779bf5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.274399 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-kube-api-access-q4trh" (OuterVolumeSpecName: "kube-api-access-q4trh") pod "54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" (UID: "54ed9bd6-44dd-42fc-8640-8e9bd2779bf5"). InnerVolumeSpecName "kube-api-access-q4trh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.303841 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" (UID: "54ed9bd6-44dd-42fc-8640-8e9bd2779bf5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.365297 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4trh\" (UniqueName: \"kubernetes.io/projected/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-kube-api-access-q4trh\") on node \"crc\" DevicePath \"\"" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.365340 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.365350 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.500061 4771 generic.go:334] "Generic (PLEG): container finished" podID="54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" containerID="c39f8ac7793311055928435ec5d285460f5c9573b24034da38d8170466d6d911" exitCode=0 Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.500113 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27nfg" event={"ID":"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5","Type":"ContainerDied","Data":"c39f8ac7793311055928435ec5d285460f5c9573b24034da38d8170466d6d911"} Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.500190 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27nfg" event={"ID":"54ed9bd6-44dd-42fc-8640-8e9bd2779bf5","Type":"ContainerDied","Data":"3097f68bf074122ea7f47ddc5d8f37cac2d974428e2a8744d33c8d3c2426a4ed"} Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.500213 4771 scope.go:117] "RemoveContainer" containerID="c39f8ac7793311055928435ec5d285460f5c9573b24034da38d8170466d6d911" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.500396 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-27nfg" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.540820 4771 scope.go:117] "RemoveContainer" containerID="561267b5b6c96e97e5b4e40da6e0637be09487241a184febd3e8805578825543" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.548118 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-27nfg"] Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.563779 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-27nfg"] Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.599551 4771 scope.go:117] "RemoveContainer" containerID="442bc533923f383c3b002785f20305583effce25fe70db590dc541469972a504" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.631663 4771 scope.go:117] "RemoveContainer" containerID="c39f8ac7793311055928435ec5d285460f5c9573b24034da38d8170466d6d911" Oct 02 10:23:02 crc kubenswrapper[4771]: E1002 10:23:02.632177 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c39f8ac7793311055928435ec5d285460f5c9573b24034da38d8170466d6d911\": container with ID starting with c39f8ac7793311055928435ec5d285460f5c9573b24034da38d8170466d6d911 not found: ID does not exist" containerID="c39f8ac7793311055928435ec5d285460f5c9573b24034da38d8170466d6d911" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.632229 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c39f8ac7793311055928435ec5d285460f5c9573b24034da38d8170466d6d911"} err="failed to get container status \"c39f8ac7793311055928435ec5d285460f5c9573b24034da38d8170466d6d911\": rpc error: code = NotFound desc = could not find container \"c39f8ac7793311055928435ec5d285460f5c9573b24034da38d8170466d6d911\": container with ID starting with c39f8ac7793311055928435ec5d285460f5c9573b24034da38d8170466d6d911 not found: ID does not exist" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.632266 4771 scope.go:117] "RemoveContainer" containerID="561267b5b6c96e97e5b4e40da6e0637be09487241a184febd3e8805578825543" Oct 02 10:23:02 crc kubenswrapper[4771]: E1002 10:23:02.632673 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"561267b5b6c96e97e5b4e40da6e0637be09487241a184febd3e8805578825543\": container with ID starting with 561267b5b6c96e97e5b4e40da6e0637be09487241a184febd3e8805578825543 not found: ID does not exist" containerID="561267b5b6c96e97e5b4e40da6e0637be09487241a184febd3e8805578825543" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.632732 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"561267b5b6c96e97e5b4e40da6e0637be09487241a184febd3e8805578825543"} err="failed to get container status \"561267b5b6c96e97e5b4e40da6e0637be09487241a184febd3e8805578825543\": rpc error: code = NotFound desc = could not find container \"561267b5b6c96e97e5b4e40da6e0637be09487241a184febd3e8805578825543\": container with ID starting with 561267b5b6c96e97e5b4e40da6e0637be09487241a184febd3e8805578825543 not found: ID does not exist" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.632777 4771 scope.go:117] "RemoveContainer" containerID="442bc533923f383c3b002785f20305583effce25fe70db590dc541469972a504" Oct 02 10:23:02 crc kubenswrapper[4771]: E1002 10:23:02.633094 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"442bc533923f383c3b002785f20305583effce25fe70db590dc541469972a504\": container with ID starting with 442bc533923f383c3b002785f20305583effce25fe70db590dc541469972a504 not found: ID does not exist" containerID="442bc533923f383c3b002785f20305583effce25fe70db590dc541469972a504" Oct 02 10:23:02 crc kubenswrapper[4771]: I1002 10:23:02.633127 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"442bc533923f383c3b002785f20305583effce25fe70db590dc541469972a504"} err="failed to get container status \"442bc533923f383c3b002785f20305583effce25fe70db590dc541469972a504\": rpc error: code = NotFound desc = could not find container \"442bc533923f383c3b002785f20305583effce25fe70db590dc541469972a504\": container with ID starting with 442bc533923f383c3b002785f20305583effce25fe70db590dc541469972a504 not found: ID does not exist" Oct 02 10:23:03 crc kubenswrapper[4771]: I1002 10:23:03.711537 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" path="/var/lib/kubelet/pods/54ed9bd6-44dd-42fc-8640-8e9bd2779bf5/volumes" Oct 02 10:23:12 crc kubenswrapper[4771]: I1002 10:23:12.147216 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:23:12 crc kubenswrapper[4771]: I1002 10:23:12.147765 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:23:12 crc kubenswrapper[4771]: I1002 10:23:12.147813 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 10:23:12 crc kubenswrapper[4771]: I1002 10:23:12.148735 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f79d932eb1d7bfadfdeb97c1d67c7b20b8fc5e7820bfc2dce0c429b41092abd6"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:23:12 crc kubenswrapper[4771]: I1002 10:23:12.148794 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://f79d932eb1d7bfadfdeb97c1d67c7b20b8fc5e7820bfc2dce0c429b41092abd6" gracePeriod=600 Oct 02 10:23:12 crc kubenswrapper[4771]: I1002 10:23:12.634671 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="f79d932eb1d7bfadfdeb97c1d67c7b20b8fc5e7820bfc2dce0c429b41092abd6" exitCode=0 Oct 02 10:23:12 crc kubenswrapper[4771]: I1002 10:23:12.634809 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"f79d932eb1d7bfadfdeb97c1d67c7b20b8fc5e7820bfc2dce0c429b41092abd6"} Oct 02 10:23:12 crc kubenswrapper[4771]: I1002 10:23:12.635084 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241"} Oct 02 10:23:12 crc kubenswrapper[4771]: I1002 10:23:12.635114 4771 scope.go:117] "RemoveContainer" containerID="c5a60cd5ae5b4d79e53cc5d0a751ef2122c64b54b623a95c882e95c9737deec2" Oct 02 10:25:12 crc kubenswrapper[4771]: I1002 10:25:12.146343 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:25:12 crc kubenswrapper[4771]: I1002 10:25:12.147037 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:25:17 crc kubenswrapper[4771]: E1002 10:25:17.653851 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c7c8501_8ae8_47d5_b947_8b9ac02be829.slice/crio-e85ee3edcadfe21735bfa9f46a426144b26dc1688dfee8a8b9c3b22da3a948cf.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c7c8501_8ae8_47d5_b947_8b9ac02be829.slice/crio-conmon-e85ee3edcadfe21735bfa9f46a426144b26dc1688dfee8a8b9c3b22da3a948cf.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:25:18 crc kubenswrapper[4771]: I1002 10:25:18.227467 4771 generic.go:334] "Generic (PLEG): container finished" podID="7c7c8501-8ae8-47d5-b947-8b9ac02be829" containerID="e85ee3edcadfe21735bfa9f46a426144b26dc1688dfee8a8b9c3b22da3a948cf" exitCode=0 Oct 02 10:25:18 crc kubenswrapper[4771]: I1002 10:25:18.227538 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" event={"ID":"7c7c8501-8ae8-47d5-b947-8b9ac02be829","Type":"ContainerDied","Data":"e85ee3edcadfe21735bfa9f46a426144b26dc1688dfee8a8b9c3b22da3a948cf"} Oct 02 10:25:19 crc kubenswrapper[4771]: I1002 10:25:19.843887 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:25:19 crc kubenswrapper[4771]: I1002 10:25:19.956303 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-libvirt-combined-ca-bundle\") pod \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " Oct 02 10:25:19 crc kubenswrapper[4771]: I1002 10:25:19.956391 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-libvirt-secret-0\") pod \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " Oct 02 10:25:19 crc kubenswrapper[4771]: I1002 10:25:19.956486 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-ssh-key\") pod \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " Oct 02 10:25:19 crc kubenswrapper[4771]: I1002 10:25:19.956699 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg29v\" (UniqueName: \"kubernetes.io/projected/7c7c8501-8ae8-47d5-b947-8b9ac02be829-kube-api-access-qg29v\") pod \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " Oct 02 10:25:19 crc kubenswrapper[4771]: I1002 10:25:19.956818 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-inventory\") pod \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\" (UID: \"7c7c8501-8ae8-47d5-b947-8b9ac02be829\") " Oct 02 10:25:19 crc kubenswrapper[4771]: I1002 10:25:19.965352 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c7c8501-8ae8-47d5-b947-8b9ac02be829-kube-api-access-qg29v" (OuterVolumeSpecName: "kube-api-access-qg29v") pod "7c7c8501-8ae8-47d5-b947-8b9ac02be829" (UID: "7c7c8501-8ae8-47d5-b947-8b9ac02be829"). InnerVolumeSpecName "kube-api-access-qg29v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:25:19 crc kubenswrapper[4771]: I1002 10:25:19.965415 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "7c7c8501-8ae8-47d5-b947-8b9ac02be829" (UID: "7c7c8501-8ae8-47d5-b947-8b9ac02be829"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.002337 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7c7c8501-8ae8-47d5-b947-8b9ac02be829" (UID: "7c7c8501-8ae8-47d5-b947-8b9ac02be829"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.003704 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-inventory" (OuterVolumeSpecName: "inventory") pod "7c7c8501-8ae8-47d5-b947-8b9ac02be829" (UID: "7c7c8501-8ae8-47d5-b947-8b9ac02be829"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.007197 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "7c7c8501-8ae8-47d5-b947-8b9ac02be829" (UID: "7c7c8501-8ae8-47d5-b947-8b9ac02be829"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.060020 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg29v\" (UniqueName: \"kubernetes.io/projected/7c7c8501-8ae8-47d5-b947-8b9ac02be829-kube-api-access-qg29v\") on node \"crc\" DevicePath \"\"" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.060077 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.060088 4771 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.060098 4771 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.060122 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c7c8501-8ae8-47d5-b947-8b9ac02be829-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.258478 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" event={"ID":"7c7c8501-8ae8-47d5-b947-8b9ac02be829","Type":"ContainerDied","Data":"cff854af87673a42acffe312f14a7da6bfd71b67e5033c1f7ed6aaf240008239"} Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.258534 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cff854af87673a42acffe312f14a7da6bfd71b67e5033c1f7ed6aaf240008239" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.258626 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.403371 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c"] Oct 02 10:25:20 crc kubenswrapper[4771]: E1002 10:25:20.404116 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" containerName="registry-server" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.404174 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" containerName="registry-server" Oct 02 10:25:20 crc kubenswrapper[4771]: E1002 10:25:20.404215 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" containerName="extract-utilities" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.404223 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" containerName="extract-utilities" Oct 02 10:25:20 crc kubenswrapper[4771]: E1002 10:25:20.404260 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" containerName="extract-content" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.404269 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" containerName="extract-content" Oct 02 10:25:20 crc kubenswrapper[4771]: E1002 10:25:20.404283 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c7c8501-8ae8-47d5-b947-8b9ac02be829" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.404291 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c7c8501-8ae8-47d5-b947-8b9ac02be829" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.404609 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="54ed9bd6-44dd-42fc-8640-8e9bd2779bf5" containerName="registry-server" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.404699 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c7c8501-8ae8-47d5-b947-8b9ac02be829" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.405846 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.408040 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.408935 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.409886 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.411423 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.415425 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.415629 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.419729 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.436429 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c"] Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.570986 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.571364 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.571398 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.571540 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.571654 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.571925 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qch49\" (UniqueName: \"kubernetes.io/projected/5242b9a6-8162-4a25-8821-5a8406cd2d86-kube-api-access-qch49\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.572045 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.572074 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.572179 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.674078 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.674219 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qch49\" (UniqueName: \"kubernetes.io/projected/5242b9a6-8162-4a25-8821-5a8406cd2d86-kube-api-access-qch49\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.674269 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.674297 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.674337 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.674366 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.674400 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.674420 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.674477 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.675484 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.679148 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.679524 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.680003 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.680738 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.682588 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.689221 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.693397 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.699531 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qch49\" (UniqueName: \"kubernetes.io/projected/5242b9a6-8162-4a25-8821-5a8406cd2d86-kube-api-access-qch49\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fq28c\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:20 crc kubenswrapper[4771]: I1002 10:25:20.739699 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:25:21 crc kubenswrapper[4771]: I1002 10:25:21.299641 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c"] Oct 02 10:25:22 crc kubenswrapper[4771]: I1002 10:25:22.288619 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" event={"ID":"5242b9a6-8162-4a25-8821-5a8406cd2d86","Type":"ContainerStarted","Data":"d9d2b6bddfd4debe19a689147140e00cc87cc33ff7da4ab696f1e290b3c21b2b"} Oct 02 10:25:22 crc kubenswrapper[4771]: I1002 10:25:22.289269 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" event={"ID":"5242b9a6-8162-4a25-8821-5a8406cd2d86","Type":"ContainerStarted","Data":"914bcc9f40bc992be33b332ae936987bbea745051e6acdf1de49664c8953f85b"} Oct 02 10:25:22 crc kubenswrapper[4771]: I1002 10:25:22.316738 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" podStartSLOduration=1.766446696 podStartE2EDuration="2.316714328s" podCreationTimestamp="2025-10-02 10:25:20 +0000 UTC" firstStartedPulling="2025-10-02 10:25:21.296588515 +0000 UTC m=+2908.944273582" lastFinishedPulling="2025-10-02 10:25:21.846856137 +0000 UTC m=+2909.494541214" observedRunningTime="2025-10-02 10:25:22.313352687 +0000 UTC m=+2909.961037794" watchObservedRunningTime="2025-10-02 10:25:22.316714328 +0000 UTC m=+2909.964399395" Oct 02 10:25:42 crc kubenswrapper[4771]: I1002 10:25:42.146376 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:25:42 crc kubenswrapper[4771]: I1002 10:25:42.146918 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:26:12 crc kubenswrapper[4771]: I1002 10:26:12.146328 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:26:12 crc kubenswrapper[4771]: I1002 10:26:12.147099 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:26:12 crc kubenswrapper[4771]: I1002 10:26:12.147203 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 10:26:12 crc kubenswrapper[4771]: I1002 10:26:12.148526 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:26:12 crc kubenswrapper[4771]: I1002 10:26:12.148630 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" gracePeriod=600 Oct 02 10:26:12 crc kubenswrapper[4771]: E1002 10:26:12.332688 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:26:12 crc kubenswrapper[4771]: I1002 10:26:12.938466 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" exitCode=0 Oct 02 10:26:12 crc kubenswrapper[4771]: I1002 10:26:12.938542 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241"} Oct 02 10:26:12 crc kubenswrapper[4771]: I1002 10:26:12.938604 4771 scope.go:117] "RemoveContainer" containerID="f79d932eb1d7bfadfdeb97c1d67c7b20b8fc5e7820bfc2dce0c429b41092abd6" Oct 02 10:26:12 crc kubenswrapper[4771]: I1002 10:26:12.940517 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:26:12 crc kubenswrapper[4771]: E1002 10:26:12.941973 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:26:23 crc kubenswrapper[4771]: I1002 10:26:23.699557 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:26:23 crc kubenswrapper[4771]: E1002 10:26:23.700675 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:26:34 crc kubenswrapper[4771]: I1002 10:26:34.681947 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:26:34 crc kubenswrapper[4771]: E1002 10:26:34.682612 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:26:45 crc kubenswrapper[4771]: I1002 10:26:45.682279 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:26:45 crc kubenswrapper[4771]: E1002 10:26:45.683705 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:26:55 crc kubenswrapper[4771]: I1002 10:26:55.764484 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g784v"] Oct 02 10:26:55 crc kubenswrapper[4771]: I1002 10:26:55.777174 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:26:55 crc kubenswrapper[4771]: I1002 10:26:55.778917 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g784v"] Oct 02 10:26:55 crc kubenswrapper[4771]: I1002 10:26:55.921913 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fcc5\" (UniqueName: \"kubernetes.io/projected/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-kube-api-access-5fcc5\") pod \"redhat-marketplace-g784v\" (UID: \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\") " pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:26:55 crc kubenswrapper[4771]: I1002 10:26:55.922082 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-catalog-content\") pod \"redhat-marketplace-g784v\" (UID: \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\") " pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:26:55 crc kubenswrapper[4771]: I1002 10:26:55.922168 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-utilities\") pod \"redhat-marketplace-g784v\" (UID: \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\") " pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:26:56 crc kubenswrapper[4771]: I1002 10:26:56.024714 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fcc5\" (UniqueName: \"kubernetes.io/projected/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-kube-api-access-5fcc5\") pod \"redhat-marketplace-g784v\" (UID: \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\") " pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:26:56 crc kubenswrapper[4771]: I1002 10:26:56.024854 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-catalog-content\") pod \"redhat-marketplace-g784v\" (UID: \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\") " pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:26:56 crc kubenswrapper[4771]: I1002 10:26:56.024908 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-utilities\") pod \"redhat-marketplace-g784v\" (UID: \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\") " pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:26:56 crc kubenswrapper[4771]: I1002 10:26:56.025551 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-catalog-content\") pod \"redhat-marketplace-g784v\" (UID: \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\") " pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:26:56 crc kubenswrapper[4771]: I1002 10:26:56.025565 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-utilities\") pod \"redhat-marketplace-g784v\" (UID: \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\") " pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:26:56 crc kubenswrapper[4771]: I1002 10:26:56.045507 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fcc5\" (UniqueName: \"kubernetes.io/projected/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-kube-api-access-5fcc5\") pod \"redhat-marketplace-g784v\" (UID: \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\") " pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:26:56 crc kubenswrapper[4771]: I1002 10:26:56.096876 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:26:56 crc kubenswrapper[4771]: I1002 10:26:56.571614 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g784v"] Oct 02 10:26:57 crc kubenswrapper[4771]: I1002 10:26:57.462046 4771 generic.go:334] "Generic (PLEG): container finished" podID="4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" containerID="df63fa377e860c81e77fd2019dd1176807f7976b2e1f0b0c95d003a7fd7039b2" exitCode=0 Oct 02 10:26:57 crc kubenswrapper[4771]: I1002 10:26:57.462191 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g784v" event={"ID":"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9","Type":"ContainerDied","Data":"df63fa377e860c81e77fd2019dd1176807f7976b2e1f0b0c95d003a7fd7039b2"} Oct 02 10:26:57 crc kubenswrapper[4771]: I1002 10:26:57.462860 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g784v" event={"ID":"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9","Type":"ContainerStarted","Data":"505134e0c70bee543deb3edb83c71eb41fbbf5e94bd9a3cefc8a330bed5c213b"} Oct 02 10:26:57 crc kubenswrapper[4771]: I1002 10:26:57.681355 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:26:57 crc kubenswrapper[4771]: E1002 10:26:57.681959 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:27:01 crc kubenswrapper[4771]: I1002 10:27:01.514777 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g784v" event={"ID":"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9","Type":"ContainerStarted","Data":"4bd4eaac4a61f2064331636c0c6a17401687cb4ec4ad28a5d3835de624a163c5"} Oct 02 10:27:02 crc kubenswrapper[4771]: I1002 10:27:02.533782 4771 generic.go:334] "Generic (PLEG): container finished" podID="4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" containerID="4bd4eaac4a61f2064331636c0c6a17401687cb4ec4ad28a5d3835de624a163c5" exitCode=0 Oct 02 10:27:02 crc kubenswrapper[4771]: I1002 10:27:02.534118 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g784v" event={"ID":"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9","Type":"ContainerDied","Data":"4bd4eaac4a61f2064331636c0c6a17401687cb4ec4ad28a5d3835de624a163c5"} Oct 02 10:27:04 crc kubenswrapper[4771]: I1002 10:27:04.561649 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g784v" event={"ID":"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9","Type":"ContainerStarted","Data":"bfcf856dbd5ebdb6ad7f509612abc20b81d5789fe78775457fe4b821cc4b5a47"} Oct 02 10:27:04 crc kubenswrapper[4771]: I1002 10:27:04.582339 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g784v" podStartSLOduration=3.752794239 podStartE2EDuration="9.582321519s" podCreationTimestamp="2025-10-02 10:26:55 +0000 UTC" firstStartedPulling="2025-10-02 10:26:57.464024831 +0000 UTC m=+3005.111709938" lastFinishedPulling="2025-10-02 10:27:03.293552151 +0000 UTC m=+3010.941237218" observedRunningTime="2025-10-02 10:27:04.577198332 +0000 UTC m=+3012.224883439" watchObservedRunningTime="2025-10-02 10:27:04.582321519 +0000 UTC m=+3012.230006586" Oct 02 10:27:06 crc kubenswrapper[4771]: I1002 10:27:06.097587 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:27:06 crc kubenswrapper[4771]: I1002 10:27:06.097911 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:27:06 crc kubenswrapper[4771]: I1002 10:27:06.147002 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:27:10 crc kubenswrapper[4771]: I1002 10:27:10.681406 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:27:10 crc kubenswrapper[4771]: E1002 10:27:10.684059 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:27:16 crc kubenswrapper[4771]: I1002 10:27:16.145281 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:27:16 crc kubenswrapper[4771]: I1002 10:27:16.205949 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g784v"] Oct 02 10:27:16 crc kubenswrapper[4771]: I1002 10:27:16.745496 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g784v" podUID="4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" containerName="registry-server" containerID="cri-o://bfcf856dbd5ebdb6ad7f509612abc20b81d5789fe78775457fe4b821cc4b5a47" gracePeriod=2 Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.261702 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.383243 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-utilities\") pod \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\" (UID: \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\") " Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.383442 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fcc5\" (UniqueName: \"kubernetes.io/projected/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-kube-api-access-5fcc5\") pod \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\" (UID: \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\") " Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.383528 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-catalog-content\") pod \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\" (UID: \"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9\") " Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.384717 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-utilities" (OuterVolumeSpecName: "utilities") pod "4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" (UID: "4e0293fb-cca9-4dbc-ba31-1e55f1e696c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.392321 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-kube-api-access-5fcc5" (OuterVolumeSpecName: "kube-api-access-5fcc5") pod "4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" (UID: "4e0293fb-cca9-4dbc-ba31-1e55f1e696c9"). InnerVolumeSpecName "kube-api-access-5fcc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.398466 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" (UID: "4e0293fb-cca9-4dbc-ba31-1e55f1e696c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.487072 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.487124 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fcc5\" (UniqueName: \"kubernetes.io/projected/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-kube-api-access-5fcc5\") on node \"crc\" DevicePath \"\"" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.487152 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.771638 4771 generic.go:334] "Generic (PLEG): container finished" podID="4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" containerID="bfcf856dbd5ebdb6ad7f509612abc20b81d5789fe78775457fe4b821cc4b5a47" exitCode=0 Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.771697 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g784v" event={"ID":"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9","Type":"ContainerDied","Data":"bfcf856dbd5ebdb6ad7f509612abc20b81d5789fe78775457fe4b821cc4b5a47"} Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.771741 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g784v" event={"ID":"4e0293fb-cca9-4dbc-ba31-1e55f1e696c9","Type":"ContainerDied","Data":"505134e0c70bee543deb3edb83c71eb41fbbf5e94bd9a3cefc8a330bed5c213b"} Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.771758 4771 scope.go:117] "RemoveContainer" containerID="bfcf856dbd5ebdb6ad7f509612abc20b81d5789fe78775457fe4b821cc4b5a47" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.771701 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g784v" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.801273 4771 scope.go:117] "RemoveContainer" containerID="4bd4eaac4a61f2064331636c0c6a17401687cb4ec4ad28a5d3835de624a163c5" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.824165 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g784v"] Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.834207 4771 scope.go:117] "RemoveContainer" containerID="df63fa377e860c81e77fd2019dd1176807f7976b2e1f0b0c95d003a7fd7039b2" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.837857 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g784v"] Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.885607 4771 scope.go:117] "RemoveContainer" containerID="bfcf856dbd5ebdb6ad7f509612abc20b81d5789fe78775457fe4b821cc4b5a47" Oct 02 10:27:17 crc kubenswrapper[4771]: E1002 10:27:17.886747 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfcf856dbd5ebdb6ad7f509612abc20b81d5789fe78775457fe4b821cc4b5a47\": container with ID starting with bfcf856dbd5ebdb6ad7f509612abc20b81d5789fe78775457fe4b821cc4b5a47 not found: ID does not exist" containerID="bfcf856dbd5ebdb6ad7f509612abc20b81d5789fe78775457fe4b821cc4b5a47" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.886796 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfcf856dbd5ebdb6ad7f509612abc20b81d5789fe78775457fe4b821cc4b5a47"} err="failed to get container status \"bfcf856dbd5ebdb6ad7f509612abc20b81d5789fe78775457fe4b821cc4b5a47\": rpc error: code = NotFound desc = could not find container \"bfcf856dbd5ebdb6ad7f509612abc20b81d5789fe78775457fe4b821cc4b5a47\": container with ID starting with bfcf856dbd5ebdb6ad7f509612abc20b81d5789fe78775457fe4b821cc4b5a47 not found: ID does not exist" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.886827 4771 scope.go:117] "RemoveContainer" containerID="4bd4eaac4a61f2064331636c0c6a17401687cb4ec4ad28a5d3835de624a163c5" Oct 02 10:27:17 crc kubenswrapper[4771]: E1002 10:27:17.887370 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bd4eaac4a61f2064331636c0c6a17401687cb4ec4ad28a5d3835de624a163c5\": container with ID starting with 4bd4eaac4a61f2064331636c0c6a17401687cb4ec4ad28a5d3835de624a163c5 not found: ID does not exist" containerID="4bd4eaac4a61f2064331636c0c6a17401687cb4ec4ad28a5d3835de624a163c5" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.887475 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bd4eaac4a61f2064331636c0c6a17401687cb4ec4ad28a5d3835de624a163c5"} err="failed to get container status \"4bd4eaac4a61f2064331636c0c6a17401687cb4ec4ad28a5d3835de624a163c5\": rpc error: code = NotFound desc = could not find container \"4bd4eaac4a61f2064331636c0c6a17401687cb4ec4ad28a5d3835de624a163c5\": container with ID starting with 4bd4eaac4a61f2064331636c0c6a17401687cb4ec4ad28a5d3835de624a163c5 not found: ID does not exist" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.887577 4771 scope.go:117] "RemoveContainer" containerID="df63fa377e860c81e77fd2019dd1176807f7976b2e1f0b0c95d003a7fd7039b2" Oct 02 10:27:17 crc kubenswrapper[4771]: E1002 10:27:17.888059 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df63fa377e860c81e77fd2019dd1176807f7976b2e1f0b0c95d003a7fd7039b2\": container with ID starting with df63fa377e860c81e77fd2019dd1176807f7976b2e1f0b0c95d003a7fd7039b2 not found: ID does not exist" containerID="df63fa377e860c81e77fd2019dd1176807f7976b2e1f0b0c95d003a7fd7039b2" Oct 02 10:27:17 crc kubenswrapper[4771]: I1002 10:27:17.888107 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df63fa377e860c81e77fd2019dd1176807f7976b2e1f0b0c95d003a7fd7039b2"} err="failed to get container status \"df63fa377e860c81e77fd2019dd1176807f7976b2e1f0b0c95d003a7fd7039b2\": rpc error: code = NotFound desc = could not find container \"df63fa377e860c81e77fd2019dd1176807f7976b2e1f0b0c95d003a7fd7039b2\": container with ID starting with df63fa377e860c81e77fd2019dd1176807f7976b2e1f0b0c95d003a7fd7039b2 not found: ID does not exist" Oct 02 10:27:19 crc kubenswrapper[4771]: I1002 10:27:19.695082 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" path="/var/lib/kubelet/pods/4e0293fb-cca9-4dbc-ba31-1e55f1e696c9/volumes" Oct 02 10:27:25 crc kubenswrapper[4771]: I1002 10:27:25.681714 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:27:25 crc kubenswrapper[4771]: E1002 10:27:25.682647 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.260884 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-brg42"] Oct 02 10:27:36 crc kubenswrapper[4771]: E1002 10:27:36.261925 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" containerName="registry-server" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.261940 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" containerName="registry-server" Oct 02 10:27:36 crc kubenswrapper[4771]: E1002 10:27:36.261962 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" containerName="extract-content" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.261972 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" containerName="extract-content" Oct 02 10:27:36 crc kubenswrapper[4771]: E1002 10:27:36.262024 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" containerName="extract-utilities" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.262033 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" containerName="extract-utilities" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.262335 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e0293fb-cca9-4dbc-ba31-1e55f1e696c9" containerName="registry-server" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.264494 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.276775 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-brg42"] Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.342311 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdjf5\" (UniqueName: \"kubernetes.io/projected/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-kube-api-access-jdjf5\") pod \"community-operators-brg42\" (UID: \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\") " pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.343070 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-utilities\") pod \"community-operators-brg42\" (UID: \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\") " pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.343153 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-catalog-content\") pod \"community-operators-brg42\" (UID: \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\") " pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.445846 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-utilities\") pod \"community-operators-brg42\" (UID: \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\") " pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.445901 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-catalog-content\") pod \"community-operators-brg42\" (UID: \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\") " pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.445946 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdjf5\" (UniqueName: \"kubernetes.io/projected/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-kube-api-access-jdjf5\") pod \"community-operators-brg42\" (UID: \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\") " pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.446993 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-utilities\") pod \"community-operators-brg42\" (UID: \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\") " pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.447019 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-catalog-content\") pod \"community-operators-brg42\" (UID: \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\") " pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.485253 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdjf5\" (UniqueName: \"kubernetes.io/projected/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-kube-api-access-jdjf5\") pod \"community-operators-brg42\" (UID: \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\") " pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:36 crc kubenswrapper[4771]: I1002 10:27:36.590190 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:37 crc kubenswrapper[4771]: I1002 10:27:37.207590 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-brg42"] Oct 02 10:27:37 crc kubenswrapper[4771]: W1002 10:27:37.213383 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7a9218e_e4ff_45a2_8476_d2a77c18ebbc.slice/crio-5e66ff96022d6062b47392ba5ea29316915c5d94aeab540c40284e674c217ebc WatchSource:0}: Error finding container 5e66ff96022d6062b47392ba5ea29316915c5d94aeab540c40284e674c217ebc: Status 404 returned error can't find the container with id 5e66ff96022d6062b47392ba5ea29316915c5d94aeab540c40284e674c217ebc Oct 02 10:27:38 crc kubenswrapper[4771]: I1002 10:27:38.030866 4771 generic.go:334] "Generic (PLEG): container finished" podID="a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" containerID="65c13ed270c78f81a24dc89c0bd6476e8df5ba214833fc65f7761dda3e793581" exitCode=0 Oct 02 10:27:38 crc kubenswrapper[4771]: I1002 10:27:38.031116 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brg42" event={"ID":"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc","Type":"ContainerDied","Data":"65c13ed270c78f81a24dc89c0bd6476e8df5ba214833fc65f7761dda3e793581"} Oct 02 10:27:38 crc kubenswrapper[4771]: I1002 10:27:38.031524 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brg42" event={"ID":"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc","Type":"ContainerStarted","Data":"5e66ff96022d6062b47392ba5ea29316915c5d94aeab540c40284e674c217ebc"} Oct 02 10:27:38 crc kubenswrapper[4771]: I1002 10:27:38.035810 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:27:40 crc kubenswrapper[4771]: I1002 10:27:40.055184 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brg42" event={"ID":"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc","Type":"ContainerStarted","Data":"7098e301bbaec58a10c26a9286cd0cb5bffa3bc5c5962a738ad5a764d1991e6c"} Oct 02 10:27:40 crc kubenswrapper[4771]: I1002 10:27:40.682564 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:27:40 crc kubenswrapper[4771]: E1002 10:27:40.682901 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:27:41 crc kubenswrapper[4771]: I1002 10:27:41.067882 4771 generic.go:334] "Generic (PLEG): container finished" podID="a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" containerID="7098e301bbaec58a10c26a9286cd0cb5bffa3bc5c5962a738ad5a764d1991e6c" exitCode=0 Oct 02 10:27:41 crc kubenswrapper[4771]: I1002 10:27:41.067957 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brg42" event={"ID":"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc","Type":"ContainerDied","Data":"7098e301bbaec58a10c26a9286cd0cb5bffa3bc5c5962a738ad5a764d1991e6c"} Oct 02 10:27:42 crc kubenswrapper[4771]: I1002 10:27:42.089970 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brg42" event={"ID":"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc","Type":"ContainerStarted","Data":"9bb0c0f17cd5cf7522b1a61fed77eda48411affb4a190486c85df8ccaa58ec06"} Oct 02 10:27:42 crc kubenswrapper[4771]: I1002 10:27:42.131425 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-brg42" podStartSLOduration=2.424455798 podStartE2EDuration="6.131395202s" podCreationTimestamp="2025-10-02 10:27:36 +0000 UTC" firstStartedPulling="2025-10-02 10:27:38.035385921 +0000 UTC m=+3045.683071018" lastFinishedPulling="2025-10-02 10:27:41.742325355 +0000 UTC m=+3049.390010422" observedRunningTime="2025-10-02 10:27:42.121018986 +0000 UTC m=+3049.768704083" watchObservedRunningTime="2025-10-02 10:27:42.131395202 +0000 UTC m=+3049.779080299" Oct 02 10:27:46 crc kubenswrapper[4771]: I1002 10:27:46.591038 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:46 crc kubenswrapper[4771]: I1002 10:27:46.591952 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:47 crc kubenswrapper[4771]: I1002 10:27:47.663577 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-brg42" podUID="a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" containerName="registry-server" probeResult="failure" output=< Oct 02 10:27:47 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:27:47 crc kubenswrapper[4771]: > Oct 02 10:27:54 crc kubenswrapper[4771]: I1002 10:27:54.682244 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:27:54 crc kubenswrapper[4771]: E1002 10:27:54.683620 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:27:56 crc kubenswrapper[4771]: I1002 10:27:56.661453 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:56 crc kubenswrapper[4771]: I1002 10:27:56.732987 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:56 crc kubenswrapper[4771]: I1002 10:27:56.903586 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-brg42"] Oct 02 10:27:58 crc kubenswrapper[4771]: I1002 10:27:58.355237 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-brg42" podUID="a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" containerName="registry-server" containerID="cri-o://9bb0c0f17cd5cf7522b1a61fed77eda48411affb4a190486c85df8ccaa58ec06" gracePeriod=2 Oct 02 10:27:58 crc kubenswrapper[4771]: I1002 10:27:58.908838 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.056696 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-utilities\") pod \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\" (UID: \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\") " Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.056824 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdjf5\" (UniqueName: \"kubernetes.io/projected/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-kube-api-access-jdjf5\") pod \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\" (UID: \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\") " Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.057044 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-catalog-content\") pod \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\" (UID: \"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc\") " Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.057703 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-utilities" (OuterVolumeSpecName: "utilities") pod "a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" (UID: "a7a9218e-e4ff-45a2-8476-d2a77c18ebbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.058980 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.065062 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-kube-api-access-jdjf5" (OuterVolumeSpecName: "kube-api-access-jdjf5") pod "a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" (UID: "a7a9218e-e4ff-45a2-8476-d2a77c18ebbc"). InnerVolumeSpecName "kube-api-access-jdjf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.131080 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" (UID: "a7a9218e-e4ff-45a2-8476-d2a77c18ebbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.161790 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdjf5\" (UniqueName: \"kubernetes.io/projected/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-kube-api-access-jdjf5\") on node \"crc\" DevicePath \"\"" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.161834 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.369817 4771 generic.go:334] "Generic (PLEG): container finished" podID="a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" containerID="9bb0c0f17cd5cf7522b1a61fed77eda48411affb4a190486c85df8ccaa58ec06" exitCode=0 Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.369865 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brg42" event={"ID":"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc","Type":"ContainerDied","Data":"9bb0c0f17cd5cf7522b1a61fed77eda48411affb4a190486c85df8ccaa58ec06"} Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.369897 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brg42" event={"ID":"a7a9218e-e4ff-45a2-8476-d2a77c18ebbc","Type":"ContainerDied","Data":"5e66ff96022d6062b47392ba5ea29316915c5d94aeab540c40284e674c217ebc"} Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.369925 4771 scope.go:117] "RemoveContainer" containerID="9bb0c0f17cd5cf7522b1a61fed77eda48411affb4a190486c85df8ccaa58ec06" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.370086 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brg42" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.396113 4771 scope.go:117] "RemoveContainer" containerID="7098e301bbaec58a10c26a9286cd0cb5bffa3bc5c5962a738ad5a764d1991e6c" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.424140 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-brg42"] Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.437538 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-brg42"] Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.439576 4771 scope.go:117] "RemoveContainer" containerID="65c13ed270c78f81a24dc89c0bd6476e8df5ba214833fc65f7761dda3e793581" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.496811 4771 scope.go:117] "RemoveContainer" containerID="9bb0c0f17cd5cf7522b1a61fed77eda48411affb4a190486c85df8ccaa58ec06" Oct 02 10:27:59 crc kubenswrapper[4771]: E1002 10:27:59.497309 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bb0c0f17cd5cf7522b1a61fed77eda48411affb4a190486c85df8ccaa58ec06\": container with ID starting with 9bb0c0f17cd5cf7522b1a61fed77eda48411affb4a190486c85df8ccaa58ec06 not found: ID does not exist" containerID="9bb0c0f17cd5cf7522b1a61fed77eda48411affb4a190486c85df8ccaa58ec06" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.497363 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bb0c0f17cd5cf7522b1a61fed77eda48411affb4a190486c85df8ccaa58ec06"} err="failed to get container status \"9bb0c0f17cd5cf7522b1a61fed77eda48411affb4a190486c85df8ccaa58ec06\": rpc error: code = NotFound desc = could not find container \"9bb0c0f17cd5cf7522b1a61fed77eda48411affb4a190486c85df8ccaa58ec06\": container with ID starting with 9bb0c0f17cd5cf7522b1a61fed77eda48411affb4a190486c85df8ccaa58ec06 not found: ID does not exist" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.497400 4771 scope.go:117] "RemoveContainer" containerID="7098e301bbaec58a10c26a9286cd0cb5bffa3bc5c5962a738ad5a764d1991e6c" Oct 02 10:27:59 crc kubenswrapper[4771]: E1002 10:27:59.497885 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7098e301bbaec58a10c26a9286cd0cb5bffa3bc5c5962a738ad5a764d1991e6c\": container with ID starting with 7098e301bbaec58a10c26a9286cd0cb5bffa3bc5c5962a738ad5a764d1991e6c not found: ID does not exist" containerID="7098e301bbaec58a10c26a9286cd0cb5bffa3bc5c5962a738ad5a764d1991e6c" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.497923 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7098e301bbaec58a10c26a9286cd0cb5bffa3bc5c5962a738ad5a764d1991e6c"} err="failed to get container status \"7098e301bbaec58a10c26a9286cd0cb5bffa3bc5c5962a738ad5a764d1991e6c\": rpc error: code = NotFound desc = could not find container \"7098e301bbaec58a10c26a9286cd0cb5bffa3bc5c5962a738ad5a764d1991e6c\": container with ID starting with 7098e301bbaec58a10c26a9286cd0cb5bffa3bc5c5962a738ad5a764d1991e6c not found: ID does not exist" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.497943 4771 scope.go:117] "RemoveContainer" containerID="65c13ed270c78f81a24dc89c0bd6476e8df5ba214833fc65f7761dda3e793581" Oct 02 10:27:59 crc kubenswrapper[4771]: E1002 10:27:59.498357 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65c13ed270c78f81a24dc89c0bd6476e8df5ba214833fc65f7761dda3e793581\": container with ID starting with 65c13ed270c78f81a24dc89c0bd6476e8df5ba214833fc65f7761dda3e793581 not found: ID does not exist" containerID="65c13ed270c78f81a24dc89c0bd6476e8df5ba214833fc65f7761dda3e793581" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.498397 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65c13ed270c78f81a24dc89c0bd6476e8df5ba214833fc65f7761dda3e793581"} err="failed to get container status \"65c13ed270c78f81a24dc89c0bd6476e8df5ba214833fc65f7761dda3e793581\": rpc error: code = NotFound desc = could not find container \"65c13ed270c78f81a24dc89c0bd6476e8df5ba214833fc65f7761dda3e793581\": container with ID starting with 65c13ed270c78f81a24dc89c0bd6476e8df5ba214833fc65f7761dda3e793581 not found: ID does not exist" Oct 02 10:27:59 crc kubenswrapper[4771]: I1002 10:27:59.693409 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" path="/var/lib/kubelet/pods/a7a9218e-e4ff-45a2-8476-d2a77c18ebbc/volumes" Oct 02 10:28:07 crc kubenswrapper[4771]: I1002 10:28:07.681828 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:28:07 crc kubenswrapper[4771]: E1002 10:28:07.683408 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:28:22 crc kubenswrapper[4771]: I1002 10:28:22.681722 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:28:22 crc kubenswrapper[4771]: E1002 10:28:22.682608 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:28:36 crc kubenswrapper[4771]: I1002 10:28:36.680780 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:28:36 crc kubenswrapper[4771]: E1002 10:28:36.681754 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:28:50 crc kubenswrapper[4771]: I1002 10:28:50.682535 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:28:50 crc kubenswrapper[4771]: E1002 10:28:50.683677 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:29:04 crc kubenswrapper[4771]: I1002 10:29:04.681583 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:29:04 crc kubenswrapper[4771]: E1002 10:29:04.682455 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:29:14 crc kubenswrapper[4771]: I1002 10:29:14.341738 4771 generic.go:334] "Generic (PLEG): container finished" podID="5242b9a6-8162-4a25-8821-5a8406cd2d86" containerID="d9d2b6bddfd4debe19a689147140e00cc87cc33ff7da4ab696f1e290b3c21b2b" exitCode=0 Oct 02 10:29:14 crc kubenswrapper[4771]: I1002 10:29:14.341926 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" event={"ID":"5242b9a6-8162-4a25-8821-5a8406cd2d86","Type":"ContainerDied","Data":"d9d2b6bddfd4debe19a689147140e00cc87cc33ff7da4ab696f1e290b3c21b2b"} Oct 02 10:29:15 crc kubenswrapper[4771]: I1002 10:29:15.906911 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.092294 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-extra-config-0\") pod \"5242b9a6-8162-4a25-8821-5a8406cd2d86\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.092488 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-migration-ssh-key-0\") pod \"5242b9a6-8162-4a25-8821-5a8406cd2d86\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.092528 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-inventory\") pod \"5242b9a6-8162-4a25-8821-5a8406cd2d86\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.092585 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qch49\" (UniqueName: \"kubernetes.io/projected/5242b9a6-8162-4a25-8821-5a8406cd2d86-kube-api-access-qch49\") pod \"5242b9a6-8162-4a25-8821-5a8406cd2d86\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.092641 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-ssh-key\") pod \"5242b9a6-8162-4a25-8821-5a8406cd2d86\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.092698 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-combined-ca-bundle\") pod \"5242b9a6-8162-4a25-8821-5a8406cd2d86\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.092771 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-cell1-compute-config-1\") pod \"5242b9a6-8162-4a25-8821-5a8406cd2d86\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.092939 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-cell1-compute-config-0\") pod \"5242b9a6-8162-4a25-8821-5a8406cd2d86\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.093000 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-migration-ssh-key-1\") pod \"5242b9a6-8162-4a25-8821-5a8406cd2d86\" (UID: \"5242b9a6-8162-4a25-8821-5a8406cd2d86\") " Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.120810 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "5242b9a6-8162-4a25-8821-5a8406cd2d86" (UID: "5242b9a6-8162-4a25-8821-5a8406cd2d86"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.152364 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5242b9a6-8162-4a25-8821-5a8406cd2d86-kube-api-access-qch49" (OuterVolumeSpecName: "kube-api-access-qch49") pod "5242b9a6-8162-4a25-8821-5a8406cd2d86" (UID: "5242b9a6-8162-4a25-8821-5a8406cd2d86"). InnerVolumeSpecName "kube-api-access-qch49". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.183934 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "5242b9a6-8162-4a25-8821-5a8406cd2d86" (UID: "5242b9a6-8162-4a25-8821-5a8406cd2d86"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.185220 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5242b9a6-8162-4a25-8821-5a8406cd2d86" (UID: "5242b9a6-8162-4a25-8821-5a8406cd2d86"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.191000 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-inventory" (OuterVolumeSpecName: "inventory") pod "5242b9a6-8162-4a25-8821-5a8406cd2d86" (UID: "5242b9a6-8162-4a25-8821-5a8406cd2d86"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.197298 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.197348 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qch49\" (UniqueName: \"kubernetes.io/projected/5242b9a6-8162-4a25-8821-5a8406cd2d86-kube-api-access-qch49\") on node \"crc\" DevicePath \"\"" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.197363 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.197375 4771 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.197389 4771 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.206721 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "5242b9a6-8162-4a25-8821-5a8406cd2d86" (UID: "5242b9a6-8162-4a25-8821-5a8406cd2d86"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.219039 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "5242b9a6-8162-4a25-8821-5a8406cd2d86" (UID: "5242b9a6-8162-4a25-8821-5a8406cd2d86"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.222772 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "5242b9a6-8162-4a25-8821-5a8406cd2d86" (UID: "5242b9a6-8162-4a25-8821-5a8406cd2d86"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.229207 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "5242b9a6-8162-4a25-8821-5a8406cd2d86" (UID: "5242b9a6-8162-4a25-8821-5a8406cd2d86"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.299070 4771 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.299101 4771 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.299111 4771 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.299120 4771 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5242b9a6-8162-4a25-8821-5a8406cd2d86-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.368635 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" event={"ID":"5242b9a6-8162-4a25-8821-5a8406cd2d86","Type":"ContainerDied","Data":"914bcc9f40bc992be33b332ae936987bbea745051e6acdf1de49664c8953f85b"} Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.368920 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="914bcc9f40bc992be33b332ae936987bbea745051e6acdf1de49664c8953f85b" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.368704 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fq28c" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.474510 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk"] Oct 02 10:29:16 crc kubenswrapper[4771]: E1002 10:29:16.474982 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" containerName="registry-server" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.475001 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" containerName="registry-server" Oct 02 10:29:16 crc kubenswrapper[4771]: E1002 10:29:16.475030 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5242b9a6-8162-4a25-8821-5a8406cd2d86" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.475036 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="5242b9a6-8162-4a25-8821-5a8406cd2d86" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 10:29:16 crc kubenswrapper[4771]: E1002 10:29:16.475058 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" containerName="extract-utilities" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.475064 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" containerName="extract-utilities" Oct 02 10:29:16 crc kubenswrapper[4771]: E1002 10:29:16.475076 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" containerName="extract-content" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.475082 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" containerName="extract-content" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.475319 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7a9218e-e4ff-45a2-8476-d2a77c18ebbc" containerName="registry-server" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.475354 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="5242b9a6-8162-4a25-8821-5a8406cd2d86" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.476092 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.478755 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.478823 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.479756 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.480328 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.480335 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.496625 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk"] Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.607121 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.607257 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.607306 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.607351 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.607409 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwx9n\" (UniqueName: \"kubernetes.io/projected/21219956-7fb8-4a13-8b59-6f55702fe548-kube-api-access-qwx9n\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.607558 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.607691 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.709986 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwx9n\" (UniqueName: \"kubernetes.io/projected/21219956-7fb8-4a13-8b59-6f55702fe548-kube-api-access-qwx9n\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.710053 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.710099 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.710234 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.710279 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.710311 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.710350 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.715389 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.715794 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.716740 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.716942 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.717644 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.718886 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.732089 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwx9n\" (UniqueName: \"kubernetes.io/projected/21219956-7fb8-4a13-8b59-6f55702fe548-kube-api-access-qwx9n\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:16 crc kubenswrapper[4771]: I1002 10:29:16.792825 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:29:17 crc kubenswrapper[4771]: I1002 10:29:17.419651 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk"] Oct 02 10:29:17 crc kubenswrapper[4771]: I1002 10:29:17.682389 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:29:17 crc kubenswrapper[4771]: E1002 10:29:17.682769 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:29:18 crc kubenswrapper[4771]: I1002 10:29:18.395588 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" event={"ID":"21219956-7fb8-4a13-8b59-6f55702fe548","Type":"ContainerStarted","Data":"367b579da16e442eeda10ca3cad4ba25f16c0cea56dc9514225a984ac8c7f9b8"} Oct 02 10:29:19 crc kubenswrapper[4771]: I1002 10:29:19.409180 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" event={"ID":"21219956-7fb8-4a13-8b59-6f55702fe548","Type":"ContainerStarted","Data":"2e754b6c60a5a23c75664f8ef2be5bf13408728ca8159ddb056191f6f87bd790"} Oct 02 10:29:19 crc kubenswrapper[4771]: I1002 10:29:19.441545 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" podStartSLOduration=2.648167083 podStartE2EDuration="3.441516742s" podCreationTimestamp="2025-10-02 10:29:16 +0000 UTC" firstStartedPulling="2025-10-02 10:29:17.415329763 +0000 UTC m=+3145.063014830" lastFinishedPulling="2025-10-02 10:29:18.208679422 +0000 UTC m=+3145.856364489" observedRunningTime="2025-10-02 10:29:19.429401926 +0000 UTC m=+3147.077087003" watchObservedRunningTime="2025-10-02 10:29:19.441516742 +0000 UTC m=+3147.089201820" Oct 02 10:29:29 crc kubenswrapper[4771]: I1002 10:29:29.682065 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:29:29 crc kubenswrapper[4771]: E1002 10:29:29.683387 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:29:40 crc kubenswrapper[4771]: I1002 10:29:40.681851 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:29:40 crc kubenswrapper[4771]: E1002 10:29:40.684283 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:29:55 crc kubenswrapper[4771]: I1002 10:29:55.681993 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:29:55 crc kubenswrapper[4771]: E1002 10:29:55.682920 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.194206 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898"] Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.198784 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.202051 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.202282 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.213120 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898"] Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.267887 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15d11e44-5df9-4282-9bd7-b3554aa7b26a-secret-volume\") pod \"collect-profiles-29323350-7j898\" (UID: \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.267939 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15d11e44-5df9-4282-9bd7-b3554aa7b26a-config-volume\") pod \"collect-profiles-29323350-7j898\" (UID: \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.268520 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhfc6\" (UniqueName: \"kubernetes.io/projected/15d11e44-5df9-4282-9bd7-b3554aa7b26a-kube-api-access-qhfc6\") pod \"collect-profiles-29323350-7j898\" (UID: \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.370326 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhfc6\" (UniqueName: \"kubernetes.io/projected/15d11e44-5df9-4282-9bd7-b3554aa7b26a-kube-api-access-qhfc6\") pod \"collect-profiles-29323350-7j898\" (UID: \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.370459 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15d11e44-5df9-4282-9bd7-b3554aa7b26a-secret-volume\") pod \"collect-profiles-29323350-7j898\" (UID: \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.370491 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15d11e44-5df9-4282-9bd7-b3554aa7b26a-config-volume\") pod \"collect-profiles-29323350-7j898\" (UID: \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.371934 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15d11e44-5df9-4282-9bd7-b3554aa7b26a-config-volume\") pod \"collect-profiles-29323350-7j898\" (UID: \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.376339 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15d11e44-5df9-4282-9bd7-b3554aa7b26a-secret-volume\") pod \"collect-profiles-29323350-7j898\" (UID: \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.395778 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhfc6\" (UniqueName: \"kubernetes.io/projected/15d11e44-5df9-4282-9bd7-b3554aa7b26a-kube-api-access-qhfc6\") pod \"collect-profiles-29323350-7j898\" (UID: \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" Oct 02 10:30:00 crc kubenswrapper[4771]: I1002 10:30:00.532221 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" Oct 02 10:30:01 crc kubenswrapper[4771]: I1002 10:30:01.016075 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898"] Oct 02 10:30:01 crc kubenswrapper[4771]: I1002 10:30:01.932766 4771 generic.go:334] "Generic (PLEG): container finished" podID="15d11e44-5df9-4282-9bd7-b3554aa7b26a" containerID="82601b0ebfffbc20917c1dc42d76d82c0dfdd4c4e401ec440e0066b50b88c533" exitCode=0 Oct 02 10:30:01 crc kubenswrapper[4771]: I1002 10:30:01.932828 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" event={"ID":"15d11e44-5df9-4282-9bd7-b3554aa7b26a","Type":"ContainerDied","Data":"82601b0ebfffbc20917c1dc42d76d82c0dfdd4c4e401ec440e0066b50b88c533"} Oct 02 10:30:01 crc kubenswrapper[4771]: I1002 10:30:01.932864 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" event={"ID":"15d11e44-5df9-4282-9bd7-b3554aa7b26a","Type":"ContainerStarted","Data":"0ce9e5aae0122df63992ef121a76bd75a336ae2a4cfad21f2f756289708f7951"} Oct 02 10:30:03 crc kubenswrapper[4771]: I1002 10:30:03.374957 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" Oct 02 10:30:03 crc kubenswrapper[4771]: I1002 10:30:03.446300 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15d11e44-5df9-4282-9bd7-b3554aa7b26a-config-volume\") pod \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\" (UID: \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\") " Oct 02 10:30:03 crc kubenswrapper[4771]: I1002 10:30:03.446384 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15d11e44-5df9-4282-9bd7-b3554aa7b26a-secret-volume\") pod \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\" (UID: \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\") " Oct 02 10:30:03 crc kubenswrapper[4771]: I1002 10:30:03.446536 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhfc6\" (UniqueName: \"kubernetes.io/projected/15d11e44-5df9-4282-9bd7-b3554aa7b26a-kube-api-access-qhfc6\") pod \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\" (UID: \"15d11e44-5df9-4282-9bd7-b3554aa7b26a\") " Oct 02 10:30:03 crc kubenswrapper[4771]: I1002 10:30:03.447705 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15d11e44-5df9-4282-9bd7-b3554aa7b26a-config-volume" (OuterVolumeSpecName: "config-volume") pod "15d11e44-5df9-4282-9bd7-b3554aa7b26a" (UID: "15d11e44-5df9-4282-9bd7-b3554aa7b26a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:30:03 crc kubenswrapper[4771]: I1002 10:30:03.452572 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15d11e44-5df9-4282-9bd7-b3554aa7b26a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "15d11e44-5df9-4282-9bd7-b3554aa7b26a" (UID: "15d11e44-5df9-4282-9bd7-b3554aa7b26a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:30:03 crc kubenswrapper[4771]: I1002 10:30:03.454531 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15d11e44-5df9-4282-9bd7-b3554aa7b26a-kube-api-access-qhfc6" (OuterVolumeSpecName: "kube-api-access-qhfc6") pod "15d11e44-5df9-4282-9bd7-b3554aa7b26a" (UID: "15d11e44-5df9-4282-9bd7-b3554aa7b26a"). InnerVolumeSpecName "kube-api-access-qhfc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:30:03 crc kubenswrapper[4771]: I1002 10:30:03.549078 4771 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15d11e44-5df9-4282-9bd7-b3554aa7b26a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:30:03 crc kubenswrapper[4771]: I1002 10:30:03.549111 4771 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15d11e44-5df9-4282-9bd7-b3554aa7b26a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:30:03 crc kubenswrapper[4771]: I1002 10:30:03.549121 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhfc6\" (UniqueName: \"kubernetes.io/projected/15d11e44-5df9-4282-9bd7-b3554aa7b26a-kube-api-access-qhfc6\") on node \"crc\" DevicePath \"\"" Oct 02 10:30:03 crc kubenswrapper[4771]: I1002 10:30:03.953973 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" event={"ID":"15d11e44-5df9-4282-9bd7-b3554aa7b26a","Type":"ContainerDied","Data":"0ce9e5aae0122df63992ef121a76bd75a336ae2a4cfad21f2f756289708f7951"} Oct 02 10:30:03 crc kubenswrapper[4771]: I1002 10:30:03.954024 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ce9e5aae0122df63992ef121a76bd75a336ae2a4cfad21f2f756289708f7951" Oct 02 10:30:03 crc kubenswrapper[4771]: I1002 10:30:03.954000 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898" Oct 02 10:30:04 crc kubenswrapper[4771]: I1002 10:30:04.470342 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz"] Oct 02 10:30:04 crc kubenswrapper[4771]: I1002 10:30:04.479478 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-cfprz"] Oct 02 10:30:05 crc kubenswrapper[4771]: I1002 10:30:05.702999 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5548296e-bb64-4fce-8cc0-26f7d7f213e3" path="/var/lib/kubelet/pods/5548296e-bb64-4fce-8cc0-26f7d7f213e3/volumes" Oct 02 10:30:08 crc kubenswrapper[4771]: I1002 10:30:08.689619 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:30:08 crc kubenswrapper[4771]: E1002 10:30:08.691845 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:30:20 crc kubenswrapper[4771]: I1002 10:30:20.682086 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:30:20 crc kubenswrapper[4771]: E1002 10:30:20.683353 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:30:35 crc kubenswrapper[4771]: I1002 10:30:35.686411 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:30:35 crc kubenswrapper[4771]: E1002 10:30:35.687128 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:30:49 crc kubenswrapper[4771]: I1002 10:30:49.682380 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:30:49 crc kubenswrapper[4771]: E1002 10:30:49.683597 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:31:01 crc kubenswrapper[4771]: I1002 10:31:01.681409 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:31:01 crc kubenswrapper[4771]: E1002 10:31:01.682405 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:31:04 crc kubenswrapper[4771]: I1002 10:31:04.848277 4771 scope.go:117] "RemoveContainer" containerID="c1594021163a791a0349db261831b382673672a437c2d41c727ee75204223ffe" Oct 02 10:31:13 crc kubenswrapper[4771]: I1002 10:31:13.690972 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:31:14 crc kubenswrapper[4771]: I1002 10:31:14.856464 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"38f585bfa8e4f2dadd77f4304ba67b1b2f4b224957e4927913f6b7f916a67b45"} Oct 02 10:32:04 crc kubenswrapper[4771]: I1002 10:32:04.511208 4771 generic.go:334] "Generic (PLEG): container finished" podID="21219956-7fb8-4a13-8b59-6f55702fe548" containerID="2e754b6c60a5a23c75664f8ef2be5bf13408728ca8159ddb056191f6f87bd790" exitCode=0 Oct 02 10:32:04 crc kubenswrapper[4771]: I1002 10:32:04.511454 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" event={"ID":"21219956-7fb8-4a13-8b59-6f55702fe548","Type":"ContainerDied","Data":"2e754b6c60a5a23c75664f8ef2be5bf13408728ca8159ddb056191f6f87bd790"} Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.092102 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.123336 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-2\") pod \"21219956-7fb8-4a13-8b59-6f55702fe548\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.123534 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwx9n\" (UniqueName: \"kubernetes.io/projected/21219956-7fb8-4a13-8b59-6f55702fe548-kube-api-access-qwx9n\") pod \"21219956-7fb8-4a13-8b59-6f55702fe548\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.123692 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-inventory\") pod \"21219956-7fb8-4a13-8b59-6f55702fe548\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.123966 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-1\") pod \"21219956-7fb8-4a13-8b59-6f55702fe548\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.124095 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-telemetry-combined-ca-bundle\") pod \"21219956-7fb8-4a13-8b59-6f55702fe548\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.124247 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ssh-key\") pod \"21219956-7fb8-4a13-8b59-6f55702fe548\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.124368 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-0\") pod \"21219956-7fb8-4a13-8b59-6f55702fe548\" (UID: \"21219956-7fb8-4a13-8b59-6f55702fe548\") " Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.133832 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21219956-7fb8-4a13-8b59-6f55702fe548-kube-api-access-qwx9n" (OuterVolumeSpecName: "kube-api-access-qwx9n") pod "21219956-7fb8-4a13-8b59-6f55702fe548" (UID: "21219956-7fb8-4a13-8b59-6f55702fe548"). InnerVolumeSpecName "kube-api-access-qwx9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.142407 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "21219956-7fb8-4a13-8b59-6f55702fe548" (UID: "21219956-7fb8-4a13-8b59-6f55702fe548"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.176302 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "21219956-7fb8-4a13-8b59-6f55702fe548" (UID: "21219956-7fb8-4a13-8b59-6f55702fe548"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.176388 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "21219956-7fb8-4a13-8b59-6f55702fe548" (UID: "21219956-7fb8-4a13-8b59-6f55702fe548"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.178233 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "21219956-7fb8-4a13-8b59-6f55702fe548" (UID: "21219956-7fb8-4a13-8b59-6f55702fe548"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.199478 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-inventory" (OuterVolumeSpecName: "inventory") pod "21219956-7fb8-4a13-8b59-6f55702fe548" (UID: "21219956-7fb8-4a13-8b59-6f55702fe548"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.201880 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "21219956-7fb8-4a13-8b59-6f55702fe548" (UID: "21219956-7fb8-4a13-8b59-6f55702fe548"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.227837 4771 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.227894 4771 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.227906 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwx9n\" (UniqueName: \"kubernetes.io/projected/21219956-7fb8-4a13-8b59-6f55702fe548-kube-api-access-qwx9n\") on node \"crc\" DevicePath \"\"" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.227921 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.227930 4771 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.227945 4771 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.227975 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21219956-7fb8-4a13-8b59-6f55702fe548-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.541528 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" event={"ID":"21219956-7fb8-4a13-8b59-6f55702fe548","Type":"ContainerDied","Data":"367b579da16e442eeda10ca3cad4ba25f16c0cea56dc9514225a984ac8c7f9b8"} Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.541564 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.541577 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="367b579da16e442eeda10ca3cad4ba25f16c0cea56dc9514225a984ac8c7f9b8" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.701062 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6"] Oct 02 10:32:06 crc kubenswrapper[4771]: E1002 10:32:06.701921 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d11e44-5df9-4282-9bd7-b3554aa7b26a" containerName="collect-profiles" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.701938 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d11e44-5df9-4282-9bd7-b3554aa7b26a" containerName="collect-profiles" Oct 02 10:32:06 crc kubenswrapper[4771]: E1002 10:32:06.701953 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21219956-7fb8-4a13-8b59-6f55702fe548" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.701961 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="21219956-7fb8-4a13-8b59-6f55702fe548" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.702353 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="15d11e44-5df9-4282-9bd7-b3554aa7b26a" containerName="collect-profiles" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.702370 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="21219956-7fb8-4a13-8b59-6f55702fe548" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.703623 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.710961 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.711242 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.711295 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.724011 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.724422 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.737232 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6"] Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.744086 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.744258 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.744367 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.744454 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.744512 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd8gg\" (UniqueName: \"kubernetes.io/projected/d6e37706-8393-4c3c-9941-3e615b320c51-kube-api-access-vd8gg\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.744601 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.744642 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.846949 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.847071 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.847138 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.847176 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd8gg\" (UniqueName: \"kubernetes.io/projected/d6e37706-8393-4c3c-9941-3e615b320c51-kube-api-access-vd8gg\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.847234 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.847269 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.847321 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.857783 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.860873 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.862851 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.863663 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.865256 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.865479 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:06 crc kubenswrapper[4771]: I1002 10:32:06.871816 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd8gg\" (UniqueName: \"kubernetes.io/projected/d6e37706-8393-4c3c-9941-3e615b320c51-kube-api-access-vd8gg\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:07 crc kubenswrapper[4771]: I1002 10:32:07.033936 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:32:07 crc kubenswrapper[4771]: I1002 10:32:07.666288 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6"] Oct 02 10:32:08 crc kubenswrapper[4771]: I1002 10:32:08.573476 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" event={"ID":"d6e37706-8393-4c3c-9941-3e615b320c51","Type":"ContainerStarted","Data":"a6660497fb2485789ac76f5ef46ffe270a4ac5717a58f61a6da93334c68d8de1"} Oct 02 10:32:09 crc kubenswrapper[4771]: I1002 10:32:09.586657 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" event={"ID":"d6e37706-8393-4c3c-9941-3e615b320c51","Type":"ContainerStarted","Data":"6f5ea7388b9b16b68d011bcb0d7ed8d840bbe3e106423c86cda6dbd5634f6cd3"} Oct 02 10:32:09 crc kubenswrapper[4771]: I1002 10:32:09.622517 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" podStartSLOduration=2.98149963 podStartE2EDuration="3.622497158s" podCreationTimestamp="2025-10-02 10:32:06 +0000 UTC" firstStartedPulling="2025-10-02 10:32:07.681928096 +0000 UTC m=+3315.329613203" lastFinishedPulling="2025-10-02 10:32:08.322925664 +0000 UTC m=+3315.970610731" observedRunningTime="2025-10-02 10:32:09.615222503 +0000 UTC m=+3317.262907570" watchObservedRunningTime="2025-10-02 10:32:09.622497158 +0000 UTC m=+3317.270182225" Oct 02 10:33:04 crc kubenswrapper[4771]: I1002 10:33:04.748889 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jvjnr"] Oct 02 10:33:04 crc kubenswrapper[4771]: I1002 10:33:04.752601 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:04 crc kubenswrapper[4771]: I1002 10:33:04.760462 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jvjnr"] Oct 02 10:33:04 crc kubenswrapper[4771]: I1002 10:33:04.797668 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zhft\" (UniqueName: \"kubernetes.io/projected/33c024c8-cdf0-4fb5-8b29-23fdf1663082-kube-api-access-8zhft\") pod \"redhat-operators-jvjnr\" (UID: \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\") " pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:04 crc kubenswrapper[4771]: I1002 10:33:04.798001 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c024c8-cdf0-4fb5-8b29-23fdf1663082-catalog-content\") pod \"redhat-operators-jvjnr\" (UID: \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\") " pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:04 crc kubenswrapper[4771]: I1002 10:33:04.798158 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c024c8-cdf0-4fb5-8b29-23fdf1663082-utilities\") pod \"redhat-operators-jvjnr\" (UID: \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\") " pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:04 crc kubenswrapper[4771]: I1002 10:33:04.899995 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c024c8-cdf0-4fb5-8b29-23fdf1663082-catalog-content\") pod \"redhat-operators-jvjnr\" (UID: \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\") " pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:04 crc kubenswrapper[4771]: I1002 10:33:04.900165 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c024c8-cdf0-4fb5-8b29-23fdf1663082-utilities\") pod \"redhat-operators-jvjnr\" (UID: \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\") " pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:04 crc kubenswrapper[4771]: I1002 10:33:04.900220 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zhft\" (UniqueName: \"kubernetes.io/projected/33c024c8-cdf0-4fb5-8b29-23fdf1663082-kube-api-access-8zhft\") pod \"redhat-operators-jvjnr\" (UID: \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\") " pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:04 crc kubenswrapper[4771]: I1002 10:33:04.900700 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c024c8-cdf0-4fb5-8b29-23fdf1663082-catalog-content\") pod \"redhat-operators-jvjnr\" (UID: \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\") " pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:04 crc kubenswrapper[4771]: I1002 10:33:04.901075 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c024c8-cdf0-4fb5-8b29-23fdf1663082-utilities\") pod \"redhat-operators-jvjnr\" (UID: \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\") " pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:04 crc kubenswrapper[4771]: I1002 10:33:04.922756 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zhft\" (UniqueName: \"kubernetes.io/projected/33c024c8-cdf0-4fb5-8b29-23fdf1663082-kube-api-access-8zhft\") pod \"redhat-operators-jvjnr\" (UID: \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\") " pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:05 crc kubenswrapper[4771]: I1002 10:33:05.076305 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:05 crc kubenswrapper[4771]: W1002 10:33:05.666517 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33c024c8_cdf0_4fb5_8b29_23fdf1663082.slice/crio-575ddcb84eba4a77a804a4092722b681b707ffa1880176ca4b98aed8c66eb8aa WatchSource:0}: Error finding container 575ddcb84eba4a77a804a4092722b681b707ffa1880176ca4b98aed8c66eb8aa: Status 404 returned error can't find the container with id 575ddcb84eba4a77a804a4092722b681b707ffa1880176ca4b98aed8c66eb8aa Oct 02 10:33:05 crc kubenswrapper[4771]: I1002 10:33:05.666893 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jvjnr"] Oct 02 10:33:06 crc kubenswrapper[4771]: I1002 10:33:06.316426 4771 generic.go:334] "Generic (PLEG): container finished" podID="33c024c8-cdf0-4fb5-8b29-23fdf1663082" containerID="46504775ebf651e22efb38e69bc0588aec1f91e0376005faba2e4d7966d3733a" exitCode=0 Oct 02 10:33:06 crc kubenswrapper[4771]: I1002 10:33:06.316748 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvjnr" event={"ID":"33c024c8-cdf0-4fb5-8b29-23fdf1663082","Type":"ContainerDied","Data":"46504775ebf651e22efb38e69bc0588aec1f91e0376005faba2e4d7966d3733a"} Oct 02 10:33:06 crc kubenswrapper[4771]: I1002 10:33:06.316836 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvjnr" event={"ID":"33c024c8-cdf0-4fb5-8b29-23fdf1663082","Type":"ContainerStarted","Data":"575ddcb84eba4a77a804a4092722b681b707ffa1880176ca4b98aed8c66eb8aa"} Oct 02 10:33:06 crc kubenswrapper[4771]: I1002 10:33:06.318862 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:33:09 crc kubenswrapper[4771]: I1002 10:33:09.378095 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvjnr" event={"ID":"33c024c8-cdf0-4fb5-8b29-23fdf1663082","Type":"ContainerStarted","Data":"0ee7fc136201c0f2ede64e05e0b08c0e6bf40b869a8d82c3836f55ed6d07bb1e"} Oct 02 10:33:17 crc kubenswrapper[4771]: I1002 10:33:17.476822 4771 generic.go:334] "Generic (PLEG): container finished" podID="33c024c8-cdf0-4fb5-8b29-23fdf1663082" containerID="0ee7fc136201c0f2ede64e05e0b08c0e6bf40b869a8d82c3836f55ed6d07bb1e" exitCode=0 Oct 02 10:33:17 crc kubenswrapper[4771]: I1002 10:33:17.476913 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvjnr" event={"ID":"33c024c8-cdf0-4fb5-8b29-23fdf1663082","Type":"ContainerDied","Data":"0ee7fc136201c0f2ede64e05e0b08c0e6bf40b869a8d82c3836f55ed6d07bb1e"} Oct 02 10:33:20 crc kubenswrapper[4771]: I1002 10:33:20.522761 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvjnr" event={"ID":"33c024c8-cdf0-4fb5-8b29-23fdf1663082","Type":"ContainerStarted","Data":"fb3a84bc9c149b7ff576adfef365d733f3389a2f42cb395c57e0617c5941ec4a"} Oct 02 10:33:20 crc kubenswrapper[4771]: I1002 10:33:20.556892 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jvjnr" podStartSLOduration=3.242995771 podStartE2EDuration="16.556869532s" podCreationTimestamp="2025-10-02 10:33:04 +0000 UTC" firstStartedPulling="2025-10-02 10:33:06.318639602 +0000 UTC m=+3373.966324669" lastFinishedPulling="2025-10-02 10:33:19.632513363 +0000 UTC m=+3387.280198430" observedRunningTime="2025-10-02 10:33:20.550096819 +0000 UTC m=+3388.197781886" watchObservedRunningTime="2025-10-02 10:33:20.556869532 +0000 UTC m=+3388.204554609" Oct 02 10:33:25 crc kubenswrapper[4771]: I1002 10:33:25.077344 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:25 crc kubenswrapper[4771]: I1002 10:33:25.079009 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:25 crc kubenswrapper[4771]: I1002 10:33:25.131200 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:25 crc kubenswrapper[4771]: I1002 10:33:25.631712 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:25 crc kubenswrapper[4771]: I1002 10:33:25.677794 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jvjnr"] Oct 02 10:33:27 crc kubenswrapper[4771]: I1002 10:33:27.605410 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jvjnr" podUID="33c024c8-cdf0-4fb5-8b29-23fdf1663082" containerName="registry-server" containerID="cri-o://fb3a84bc9c149b7ff576adfef365d733f3389a2f42cb395c57e0617c5941ec4a" gracePeriod=2 Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.383989 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.539826 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zhft\" (UniqueName: \"kubernetes.io/projected/33c024c8-cdf0-4fb5-8b29-23fdf1663082-kube-api-access-8zhft\") pod \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\" (UID: \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\") " Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.540400 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c024c8-cdf0-4fb5-8b29-23fdf1663082-utilities\") pod \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\" (UID: \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\") " Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.540920 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c024c8-cdf0-4fb5-8b29-23fdf1663082-catalog-content\") pod \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\" (UID: \"33c024c8-cdf0-4fb5-8b29-23fdf1663082\") " Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.540964 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33c024c8-cdf0-4fb5-8b29-23fdf1663082-utilities" (OuterVolumeSpecName: "utilities") pod "33c024c8-cdf0-4fb5-8b29-23fdf1663082" (UID: "33c024c8-cdf0-4fb5-8b29-23fdf1663082"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.542537 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33c024c8-cdf0-4fb5-8b29-23fdf1663082-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.547952 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33c024c8-cdf0-4fb5-8b29-23fdf1663082-kube-api-access-8zhft" (OuterVolumeSpecName: "kube-api-access-8zhft") pod "33c024c8-cdf0-4fb5-8b29-23fdf1663082" (UID: "33c024c8-cdf0-4fb5-8b29-23fdf1663082"). InnerVolumeSpecName "kube-api-access-8zhft". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.619811 4771 generic.go:334] "Generic (PLEG): container finished" podID="33c024c8-cdf0-4fb5-8b29-23fdf1663082" containerID="fb3a84bc9c149b7ff576adfef365d733f3389a2f42cb395c57e0617c5941ec4a" exitCode=0 Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.619857 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvjnr" event={"ID":"33c024c8-cdf0-4fb5-8b29-23fdf1663082","Type":"ContainerDied","Data":"fb3a84bc9c149b7ff576adfef365d733f3389a2f42cb395c57e0617c5941ec4a"} Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.619885 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvjnr" event={"ID":"33c024c8-cdf0-4fb5-8b29-23fdf1663082","Type":"ContainerDied","Data":"575ddcb84eba4a77a804a4092722b681b707ffa1880176ca4b98aed8c66eb8aa"} Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.619904 4771 scope.go:117] "RemoveContainer" containerID="fb3a84bc9c149b7ff576adfef365d733f3389a2f42cb395c57e0617c5941ec4a" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.619950 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvjnr" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.631985 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33c024c8-cdf0-4fb5-8b29-23fdf1663082-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33c024c8-cdf0-4fb5-8b29-23fdf1663082" (UID: "33c024c8-cdf0-4fb5-8b29-23fdf1663082"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.644806 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zhft\" (UniqueName: \"kubernetes.io/projected/33c024c8-cdf0-4fb5-8b29-23fdf1663082-kube-api-access-8zhft\") on node \"crc\" DevicePath \"\"" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.644836 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33c024c8-cdf0-4fb5-8b29-23fdf1663082-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.650725 4771 scope.go:117] "RemoveContainer" containerID="0ee7fc136201c0f2ede64e05e0b08c0e6bf40b869a8d82c3836f55ed6d07bb1e" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.699107 4771 scope.go:117] "RemoveContainer" containerID="46504775ebf651e22efb38e69bc0588aec1f91e0376005faba2e4d7966d3733a" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.732775 4771 scope.go:117] "RemoveContainer" containerID="fb3a84bc9c149b7ff576adfef365d733f3389a2f42cb395c57e0617c5941ec4a" Oct 02 10:33:28 crc kubenswrapper[4771]: E1002 10:33:28.733106 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb3a84bc9c149b7ff576adfef365d733f3389a2f42cb395c57e0617c5941ec4a\": container with ID starting with fb3a84bc9c149b7ff576adfef365d733f3389a2f42cb395c57e0617c5941ec4a not found: ID does not exist" containerID="fb3a84bc9c149b7ff576adfef365d733f3389a2f42cb395c57e0617c5941ec4a" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.733155 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb3a84bc9c149b7ff576adfef365d733f3389a2f42cb395c57e0617c5941ec4a"} err="failed to get container status \"fb3a84bc9c149b7ff576adfef365d733f3389a2f42cb395c57e0617c5941ec4a\": rpc error: code = NotFound desc = could not find container \"fb3a84bc9c149b7ff576adfef365d733f3389a2f42cb395c57e0617c5941ec4a\": container with ID starting with fb3a84bc9c149b7ff576adfef365d733f3389a2f42cb395c57e0617c5941ec4a not found: ID does not exist" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.733177 4771 scope.go:117] "RemoveContainer" containerID="0ee7fc136201c0f2ede64e05e0b08c0e6bf40b869a8d82c3836f55ed6d07bb1e" Oct 02 10:33:28 crc kubenswrapper[4771]: E1002 10:33:28.733432 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ee7fc136201c0f2ede64e05e0b08c0e6bf40b869a8d82c3836f55ed6d07bb1e\": container with ID starting with 0ee7fc136201c0f2ede64e05e0b08c0e6bf40b869a8d82c3836f55ed6d07bb1e not found: ID does not exist" containerID="0ee7fc136201c0f2ede64e05e0b08c0e6bf40b869a8d82c3836f55ed6d07bb1e" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.733452 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ee7fc136201c0f2ede64e05e0b08c0e6bf40b869a8d82c3836f55ed6d07bb1e"} err="failed to get container status \"0ee7fc136201c0f2ede64e05e0b08c0e6bf40b869a8d82c3836f55ed6d07bb1e\": rpc error: code = NotFound desc = could not find container \"0ee7fc136201c0f2ede64e05e0b08c0e6bf40b869a8d82c3836f55ed6d07bb1e\": container with ID starting with 0ee7fc136201c0f2ede64e05e0b08c0e6bf40b869a8d82c3836f55ed6d07bb1e not found: ID does not exist" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.733465 4771 scope.go:117] "RemoveContainer" containerID="46504775ebf651e22efb38e69bc0588aec1f91e0376005faba2e4d7966d3733a" Oct 02 10:33:28 crc kubenswrapper[4771]: E1002 10:33:28.733736 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46504775ebf651e22efb38e69bc0588aec1f91e0376005faba2e4d7966d3733a\": container with ID starting with 46504775ebf651e22efb38e69bc0588aec1f91e0376005faba2e4d7966d3733a not found: ID does not exist" containerID="46504775ebf651e22efb38e69bc0588aec1f91e0376005faba2e4d7966d3733a" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.733756 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46504775ebf651e22efb38e69bc0588aec1f91e0376005faba2e4d7966d3733a"} err="failed to get container status \"46504775ebf651e22efb38e69bc0588aec1f91e0376005faba2e4d7966d3733a\": rpc error: code = NotFound desc = could not find container \"46504775ebf651e22efb38e69bc0588aec1f91e0376005faba2e4d7966d3733a\": container with ID starting with 46504775ebf651e22efb38e69bc0588aec1f91e0376005faba2e4d7966d3733a not found: ID does not exist" Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.965372 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jvjnr"] Oct 02 10:33:28 crc kubenswrapper[4771]: I1002 10:33:28.982657 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jvjnr"] Oct 02 10:33:29 crc kubenswrapper[4771]: I1002 10:33:29.694680 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33c024c8-cdf0-4fb5-8b29-23fdf1663082" path="/var/lib/kubelet/pods/33c024c8-cdf0-4fb5-8b29-23fdf1663082/volumes" Oct 02 10:33:42 crc kubenswrapper[4771]: I1002 10:33:42.146060 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:33:42 crc kubenswrapper[4771]: I1002 10:33:42.147082 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.471827 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-trjw8"] Oct 02 10:34:02 crc kubenswrapper[4771]: E1002 10:34:02.473266 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33c024c8-cdf0-4fb5-8b29-23fdf1663082" containerName="extract-content" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.473282 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="33c024c8-cdf0-4fb5-8b29-23fdf1663082" containerName="extract-content" Oct 02 10:34:02 crc kubenswrapper[4771]: E1002 10:34:02.473349 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33c024c8-cdf0-4fb5-8b29-23fdf1663082" containerName="extract-utilities" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.473356 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="33c024c8-cdf0-4fb5-8b29-23fdf1663082" containerName="extract-utilities" Oct 02 10:34:02 crc kubenswrapper[4771]: E1002 10:34:02.473406 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33c024c8-cdf0-4fb5-8b29-23fdf1663082" containerName="registry-server" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.473412 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="33c024c8-cdf0-4fb5-8b29-23fdf1663082" containerName="registry-server" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.473658 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="33c024c8-cdf0-4fb5-8b29-23fdf1663082" containerName="registry-server" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.475954 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.490277 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-trjw8"] Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.614827 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2229d1b-34e1-4ea3-a673-7b684b2d269b-utilities\") pod \"certified-operators-trjw8\" (UID: \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\") " pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.615261 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2229d1b-34e1-4ea3-a673-7b684b2d269b-catalog-content\") pod \"certified-operators-trjw8\" (UID: \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\") " pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.615498 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd2tj\" (UniqueName: \"kubernetes.io/projected/d2229d1b-34e1-4ea3-a673-7b684b2d269b-kube-api-access-sd2tj\") pod \"certified-operators-trjw8\" (UID: \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\") " pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.717285 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2229d1b-34e1-4ea3-a673-7b684b2d269b-catalog-content\") pod \"certified-operators-trjw8\" (UID: \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\") " pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.717369 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd2tj\" (UniqueName: \"kubernetes.io/projected/d2229d1b-34e1-4ea3-a673-7b684b2d269b-kube-api-access-sd2tj\") pod \"certified-operators-trjw8\" (UID: \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\") " pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.717474 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2229d1b-34e1-4ea3-a673-7b684b2d269b-utilities\") pod \"certified-operators-trjw8\" (UID: \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\") " pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.718015 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2229d1b-34e1-4ea3-a673-7b684b2d269b-utilities\") pod \"certified-operators-trjw8\" (UID: \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\") " pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.718231 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2229d1b-34e1-4ea3-a673-7b684b2d269b-catalog-content\") pod \"certified-operators-trjw8\" (UID: \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\") " pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.739861 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd2tj\" (UniqueName: \"kubernetes.io/projected/d2229d1b-34e1-4ea3-a673-7b684b2d269b-kube-api-access-sd2tj\") pod \"certified-operators-trjw8\" (UID: \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\") " pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:02 crc kubenswrapper[4771]: I1002 10:34:02.823721 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:03 crc kubenswrapper[4771]: I1002 10:34:03.469428 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-trjw8"] Oct 02 10:34:04 crc kubenswrapper[4771]: I1002 10:34:04.069041 4771 generic.go:334] "Generic (PLEG): container finished" podID="d2229d1b-34e1-4ea3-a673-7b684b2d269b" containerID="f74407febec187c1fe9521a0d3abebcc8bb8a838d0d9d76f2b0067ee43cbfa73" exitCode=0 Oct 02 10:34:04 crc kubenswrapper[4771]: I1002 10:34:04.069164 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trjw8" event={"ID":"d2229d1b-34e1-4ea3-a673-7b684b2d269b","Type":"ContainerDied","Data":"f74407febec187c1fe9521a0d3abebcc8bb8a838d0d9d76f2b0067ee43cbfa73"} Oct 02 10:34:04 crc kubenswrapper[4771]: I1002 10:34:04.069440 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trjw8" event={"ID":"d2229d1b-34e1-4ea3-a673-7b684b2d269b","Type":"ContainerStarted","Data":"6d5e76ada7e5f49ea244bfccd6ef302199cb48725ebb2778ccda9101c715276d"} Oct 02 10:34:06 crc kubenswrapper[4771]: I1002 10:34:06.096803 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trjw8" event={"ID":"d2229d1b-34e1-4ea3-a673-7b684b2d269b","Type":"ContainerStarted","Data":"96ff20de5a237a4343b0eab0e8d7de198901c5081cfc2d1fdc3cacdc3d207dcb"} Oct 02 10:34:09 crc kubenswrapper[4771]: I1002 10:34:09.143979 4771 generic.go:334] "Generic (PLEG): container finished" podID="d2229d1b-34e1-4ea3-a673-7b684b2d269b" containerID="96ff20de5a237a4343b0eab0e8d7de198901c5081cfc2d1fdc3cacdc3d207dcb" exitCode=0 Oct 02 10:34:09 crc kubenswrapper[4771]: I1002 10:34:09.144066 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trjw8" event={"ID":"d2229d1b-34e1-4ea3-a673-7b684b2d269b","Type":"ContainerDied","Data":"96ff20de5a237a4343b0eab0e8d7de198901c5081cfc2d1fdc3cacdc3d207dcb"} Oct 02 10:34:11 crc kubenswrapper[4771]: I1002 10:34:11.167749 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trjw8" event={"ID":"d2229d1b-34e1-4ea3-a673-7b684b2d269b","Type":"ContainerStarted","Data":"0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196"} Oct 02 10:34:11 crc kubenswrapper[4771]: I1002 10:34:11.220921 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-trjw8" podStartSLOduration=3.353429644 podStartE2EDuration="9.220884925s" podCreationTimestamp="2025-10-02 10:34:02 +0000 UTC" firstStartedPulling="2025-10-02 10:34:04.070877799 +0000 UTC m=+3431.718562886" lastFinishedPulling="2025-10-02 10:34:09.93833311 +0000 UTC m=+3437.586018167" observedRunningTime="2025-10-02 10:34:11.191310197 +0000 UTC m=+3438.838995264" watchObservedRunningTime="2025-10-02 10:34:11.220884925 +0000 UTC m=+3438.868570062" Oct 02 10:34:12 crc kubenswrapper[4771]: I1002 10:34:12.146099 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:34:12 crc kubenswrapper[4771]: I1002 10:34:12.146456 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:34:12 crc kubenswrapper[4771]: I1002 10:34:12.824604 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:12 crc kubenswrapper[4771]: I1002 10:34:12.824646 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:12 crc kubenswrapper[4771]: I1002 10:34:12.871008 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:22 crc kubenswrapper[4771]: I1002 10:34:22.885406 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:22 crc kubenswrapper[4771]: I1002 10:34:22.935770 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-trjw8"] Oct 02 10:34:23 crc kubenswrapper[4771]: I1002 10:34:23.319080 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-trjw8" podUID="d2229d1b-34e1-4ea3-a673-7b684b2d269b" containerName="registry-server" containerID="cri-o://0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196" gracePeriod=2 Oct 02 10:34:23 crc kubenswrapper[4771]: E1002 10:34:23.611869 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2229d1b_34e1_4ea3_a673_7b684b2d269b.slice/crio-0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:34:23 crc kubenswrapper[4771]: I1002 10:34:23.861842 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.042491 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sd2tj\" (UniqueName: \"kubernetes.io/projected/d2229d1b-34e1-4ea3-a673-7b684b2d269b-kube-api-access-sd2tj\") pod \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\" (UID: \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\") " Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.042544 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2229d1b-34e1-4ea3-a673-7b684b2d269b-utilities\") pod \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\" (UID: \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\") " Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.042806 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2229d1b-34e1-4ea3-a673-7b684b2d269b-catalog-content\") pod \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\" (UID: \"d2229d1b-34e1-4ea3-a673-7b684b2d269b\") " Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.043276 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2229d1b-34e1-4ea3-a673-7b684b2d269b-utilities" (OuterVolumeSpecName: "utilities") pod "d2229d1b-34e1-4ea3-a673-7b684b2d269b" (UID: "d2229d1b-34e1-4ea3-a673-7b684b2d269b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.043587 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2229d1b-34e1-4ea3-a673-7b684b2d269b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.051702 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2229d1b-34e1-4ea3-a673-7b684b2d269b-kube-api-access-sd2tj" (OuterVolumeSpecName: "kube-api-access-sd2tj") pod "d2229d1b-34e1-4ea3-a673-7b684b2d269b" (UID: "d2229d1b-34e1-4ea3-a673-7b684b2d269b"). InnerVolumeSpecName "kube-api-access-sd2tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.097027 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2229d1b-34e1-4ea3-a673-7b684b2d269b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2229d1b-34e1-4ea3-a673-7b684b2d269b" (UID: "d2229d1b-34e1-4ea3-a673-7b684b2d269b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.145500 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sd2tj\" (UniqueName: \"kubernetes.io/projected/d2229d1b-34e1-4ea3-a673-7b684b2d269b-kube-api-access-sd2tj\") on node \"crc\" DevicePath \"\"" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.145554 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2229d1b-34e1-4ea3-a673-7b684b2d269b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.329566 4771 generic.go:334] "Generic (PLEG): container finished" podID="d2229d1b-34e1-4ea3-a673-7b684b2d269b" containerID="0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196" exitCode=0 Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.329620 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-trjw8" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.329624 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trjw8" event={"ID":"d2229d1b-34e1-4ea3-a673-7b684b2d269b","Type":"ContainerDied","Data":"0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196"} Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.329694 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trjw8" event={"ID":"d2229d1b-34e1-4ea3-a673-7b684b2d269b","Type":"ContainerDied","Data":"6d5e76ada7e5f49ea244bfccd6ef302199cb48725ebb2778ccda9101c715276d"} Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.329717 4771 scope.go:117] "RemoveContainer" containerID="0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.360563 4771 scope.go:117] "RemoveContainer" containerID="96ff20de5a237a4343b0eab0e8d7de198901c5081cfc2d1fdc3cacdc3d207dcb" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.365652 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-trjw8"] Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.378441 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-trjw8"] Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.384567 4771 scope.go:117] "RemoveContainer" containerID="f74407febec187c1fe9521a0d3abebcc8bb8a838d0d9d76f2b0067ee43cbfa73" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.440634 4771 scope.go:117] "RemoveContainer" containerID="0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196" Oct 02 10:34:24 crc kubenswrapper[4771]: E1002 10:34:24.441087 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196\": container with ID starting with 0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196 not found: ID does not exist" containerID="0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.441121 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196"} err="failed to get container status \"0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196\": rpc error: code = NotFound desc = could not find container \"0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196\": container with ID starting with 0d1b931edad334ebcdf76911282534279974d91465df1c51e0391da821fec196 not found: ID does not exist" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.441172 4771 scope.go:117] "RemoveContainer" containerID="96ff20de5a237a4343b0eab0e8d7de198901c5081cfc2d1fdc3cacdc3d207dcb" Oct 02 10:34:24 crc kubenswrapper[4771]: E1002 10:34:24.441754 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96ff20de5a237a4343b0eab0e8d7de198901c5081cfc2d1fdc3cacdc3d207dcb\": container with ID starting with 96ff20de5a237a4343b0eab0e8d7de198901c5081cfc2d1fdc3cacdc3d207dcb not found: ID does not exist" containerID="96ff20de5a237a4343b0eab0e8d7de198901c5081cfc2d1fdc3cacdc3d207dcb" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.441777 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96ff20de5a237a4343b0eab0e8d7de198901c5081cfc2d1fdc3cacdc3d207dcb"} err="failed to get container status \"96ff20de5a237a4343b0eab0e8d7de198901c5081cfc2d1fdc3cacdc3d207dcb\": rpc error: code = NotFound desc = could not find container \"96ff20de5a237a4343b0eab0e8d7de198901c5081cfc2d1fdc3cacdc3d207dcb\": container with ID starting with 96ff20de5a237a4343b0eab0e8d7de198901c5081cfc2d1fdc3cacdc3d207dcb not found: ID does not exist" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.441791 4771 scope.go:117] "RemoveContainer" containerID="f74407febec187c1fe9521a0d3abebcc8bb8a838d0d9d76f2b0067ee43cbfa73" Oct 02 10:34:24 crc kubenswrapper[4771]: E1002 10:34:24.442061 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f74407febec187c1fe9521a0d3abebcc8bb8a838d0d9d76f2b0067ee43cbfa73\": container with ID starting with f74407febec187c1fe9521a0d3abebcc8bb8a838d0d9d76f2b0067ee43cbfa73 not found: ID does not exist" containerID="f74407febec187c1fe9521a0d3abebcc8bb8a838d0d9d76f2b0067ee43cbfa73" Oct 02 10:34:24 crc kubenswrapper[4771]: I1002 10:34:24.442081 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f74407febec187c1fe9521a0d3abebcc8bb8a838d0d9d76f2b0067ee43cbfa73"} err="failed to get container status \"f74407febec187c1fe9521a0d3abebcc8bb8a838d0d9d76f2b0067ee43cbfa73\": rpc error: code = NotFound desc = could not find container \"f74407febec187c1fe9521a0d3abebcc8bb8a838d0d9d76f2b0067ee43cbfa73\": container with ID starting with f74407febec187c1fe9521a0d3abebcc8bb8a838d0d9d76f2b0067ee43cbfa73 not found: ID does not exist" Oct 02 10:34:25 crc kubenswrapper[4771]: I1002 10:34:25.697901 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2229d1b-34e1-4ea3-a673-7b684b2d269b" path="/var/lib/kubelet/pods/d2229d1b-34e1-4ea3-a673-7b684b2d269b/volumes" Oct 02 10:34:37 crc kubenswrapper[4771]: I1002 10:34:37.471285 4771 generic.go:334] "Generic (PLEG): container finished" podID="d6e37706-8393-4c3c-9941-3e615b320c51" containerID="6f5ea7388b9b16b68d011bcb0d7ed8d840bbe3e106423c86cda6dbd5634f6cd3" exitCode=0 Oct 02 10:34:37 crc kubenswrapper[4771]: I1002 10:34:37.471354 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" event={"ID":"d6e37706-8393-4c3c-9941-3e615b320c51","Type":"ContainerDied","Data":"6f5ea7388b9b16b68d011bcb0d7ed8d840bbe3e106423c86cda6dbd5634f6cd3"} Oct 02 10:34:38 crc kubenswrapper[4771]: I1002 10:34:38.994394 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.147694 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-1\") pod \"d6e37706-8393-4c3c-9941-3e615b320c51\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.147755 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd8gg\" (UniqueName: \"kubernetes.io/projected/d6e37706-8393-4c3c-9941-3e615b320c51-kube-api-access-vd8gg\") pod \"d6e37706-8393-4c3c-9941-3e615b320c51\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.147813 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-2\") pod \"d6e37706-8393-4c3c-9941-3e615b320c51\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.147838 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-0\") pod \"d6e37706-8393-4c3c-9941-3e615b320c51\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.147913 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ssh-key\") pod \"d6e37706-8393-4c3c-9941-3e615b320c51\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.147937 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-telemetry-power-monitoring-combined-ca-bundle\") pod \"d6e37706-8393-4c3c-9941-3e615b320c51\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.148044 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-inventory\") pod \"d6e37706-8393-4c3c-9941-3e615b320c51\" (UID: \"d6e37706-8393-4c3c-9941-3e615b320c51\") " Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.163486 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "d6e37706-8393-4c3c-9941-3e615b320c51" (UID: "d6e37706-8393-4c3c-9941-3e615b320c51"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.165812 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6e37706-8393-4c3c-9941-3e615b320c51-kube-api-access-vd8gg" (OuterVolumeSpecName: "kube-api-access-vd8gg") pod "d6e37706-8393-4c3c-9941-3e615b320c51" (UID: "d6e37706-8393-4c3c-9941-3e615b320c51"). InnerVolumeSpecName "kube-api-access-vd8gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.182770 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "d6e37706-8393-4c3c-9941-3e615b320c51" (UID: "d6e37706-8393-4c3c-9941-3e615b320c51"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.188432 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-inventory" (OuterVolumeSpecName: "inventory") pod "d6e37706-8393-4c3c-9941-3e615b320c51" (UID: "d6e37706-8393-4c3c-9941-3e615b320c51"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.190942 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d6e37706-8393-4c3c-9941-3e615b320c51" (UID: "d6e37706-8393-4c3c-9941-3e615b320c51"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.196367 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "d6e37706-8393-4c3c-9941-3e615b320c51" (UID: "d6e37706-8393-4c3c-9941-3e615b320c51"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.198771 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "d6e37706-8393-4c3c-9941-3e615b320c51" (UID: "d6e37706-8393-4c3c-9941-3e615b320c51"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.251644 4771 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.251681 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd8gg\" (UniqueName: \"kubernetes.io/projected/d6e37706-8393-4c3c-9941-3e615b320c51-kube-api-access-vd8gg\") on node \"crc\" DevicePath \"\"" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.251690 4771 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.251699 4771 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.251708 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.251718 4771 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.251729 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6e37706-8393-4c3c-9941-3e615b320c51-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.500916 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" event={"ID":"d6e37706-8393-4c3c-9941-3e615b320c51","Type":"ContainerDied","Data":"a6660497fb2485789ac76f5ef46ffe270a4ac5717a58f61a6da93334c68d8de1"} Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.501276 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6660497fb2485789ac76f5ef46ffe270a4ac5717a58f61a6da93334c68d8de1" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.500956 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.600132 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2"] Oct 02 10:34:39 crc kubenswrapper[4771]: E1002 10:34:39.600640 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e37706-8393-4c3c-9941-3e615b320c51" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.600661 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e37706-8393-4c3c-9941-3e615b320c51" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Oct 02 10:34:39 crc kubenswrapper[4771]: E1002 10:34:39.600692 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2229d1b-34e1-4ea3-a673-7b684b2d269b" containerName="registry-server" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.600699 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2229d1b-34e1-4ea3-a673-7b684b2d269b" containerName="registry-server" Oct 02 10:34:39 crc kubenswrapper[4771]: E1002 10:34:39.600717 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2229d1b-34e1-4ea3-a673-7b684b2d269b" containerName="extract-utilities" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.600723 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2229d1b-34e1-4ea3-a673-7b684b2d269b" containerName="extract-utilities" Oct 02 10:34:39 crc kubenswrapper[4771]: E1002 10:34:39.600736 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2229d1b-34e1-4ea3-a673-7b684b2d269b" containerName="extract-content" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.600741 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2229d1b-34e1-4ea3-a673-7b684b2d269b" containerName="extract-content" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.600983 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2229d1b-34e1-4ea3-a673-7b684b2d269b" containerName="registry-server" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.601020 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e37706-8393-4c3c-9941-3e615b320c51" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.601848 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.609892 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-wxph7" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.609921 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.609928 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.610059 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.610255 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.618687 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2"] Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.763609 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.763657 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tf59\" (UniqueName: \"kubernetes.io/projected/db35f8f2-75a6-4c65-a44b-09230c27f953-kube-api-access-7tf59\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.763802 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.763846 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.764270 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.866684 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.866752 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.866893 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.866961 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.866988 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tf59\" (UniqueName: \"kubernetes.io/projected/db35f8f2-75a6-4c65-a44b-09230c27f953-kube-api-access-7tf59\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.872462 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.872564 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.875493 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.883845 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.887425 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tf59\" (UniqueName: \"kubernetes.io/projected/db35f8f2-75a6-4c65-a44b-09230c27f953-kube-api-access-7tf59\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pczb2\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:39 crc kubenswrapper[4771]: I1002 10:34:39.921665 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:34:40 crc kubenswrapper[4771]: I1002 10:34:40.457917 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2"] Oct 02 10:34:40 crc kubenswrapper[4771]: I1002 10:34:40.522663 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" event={"ID":"db35f8f2-75a6-4c65-a44b-09230c27f953","Type":"ContainerStarted","Data":"83bc8b240a061d77ab7f6bbe1935030cee17c42548e1a45f7d87ea7db2ff8257"} Oct 02 10:34:41 crc kubenswrapper[4771]: I1002 10:34:41.535410 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" event={"ID":"db35f8f2-75a6-4c65-a44b-09230c27f953","Type":"ContainerStarted","Data":"6eddf0c786aad6a19337ef7aadab13933dfbf705dad5c9a6f455d35da559b0d2"} Oct 02 10:34:41 crc kubenswrapper[4771]: I1002 10:34:41.557003 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" podStartSLOduration=1.892581183 podStartE2EDuration="2.556975494s" podCreationTimestamp="2025-10-02 10:34:39 +0000 UTC" firstStartedPulling="2025-10-02 10:34:40.467678169 +0000 UTC m=+3468.115363236" lastFinishedPulling="2025-10-02 10:34:41.13207248 +0000 UTC m=+3468.779757547" observedRunningTime="2025-10-02 10:34:41.553354962 +0000 UTC m=+3469.201040059" watchObservedRunningTime="2025-10-02 10:34:41.556975494 +0000 UTC m=+3469.204660601" Oct 02 10:34:42 crc kubenswrapper[4771]: I1002 10:34:42.146994 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:34:42 crc kubenswrapper[4771]: I1002 10:34:42.147347 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:34:42 crc kubenswrapper[4771]: I1002 10:34:42.147403 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 10:34:42 crc kubenswrapper[4771]: I1002 10:34:42.148575 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"38f585bfa8e4f2dadd77f4304ba67b1b2f4b224957e4927913f6b7f916a67b45"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:34:42 crc kubenswrapper[4771]: I1002 10:34:42.148635 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://38f585bfa8e4f2dadd77f4304ba67b1b2f4b224957e4927913f6b7f916a67b45" gracePeriod=600 Oct 02 10:34:42 crc kubenswrapper[4771]: I1002 10:34:42.562329 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="38f585bfa8e4f2dadd77f4304ba67b1b2f4b224957e4927913f6b7f916a67b45" exitCode=0 Oct 02 10:34:42 crc kubenswrapper[4771]: I1002 10:34:42.562605 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"38f585bfa8e4f2dadd77f4304ba67b1b2f4b224957e4927913f6b7f916a67b45"} Oct 02 10:34:42 crc kubenswrapper[4771]: I1002 10:34:42.562728 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf"} Oct 02 10:34:42 crc kubenswrapper[4771]: I1002 10:34:42.562756 4771 scope.go:117] "RemoveContainer" containerID="528c79b48379ab584c85499c7a525dde93b699e95a6740d9f5f45d5b55447241" Oct 02 10:34:59 crc kubenswrapper[4771]: I1002 10:34:59.828759 4771 generic.go:334] "Generic (PLEG): container finished" podID="db35f8f2-75a6-4c65-a44b-09230c27f953" containerID="6eddf0c786aad6a19337ef7aadab13933dfbf705dad5c9a6f455d35da559b0d2" exitCode=0 Oct 02 10:34:59 crc kubenswrapper[4771]: I1002 10:34:59.828837 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" event={"ID":"db35f8f2-75a6-4c65-a44b-09230c27f953","Type":"ContainerDied","Data":"6eddf0c786aad6a19337ef7aadab13933dfbf705dad5c9a6f455d35da559b0d2"} Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.319031 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.471979 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-logging-compute-config-data-0\") pod \"db35f8f2-75a6-4c65-a44b-09230c27f953\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.472088 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tf59\" (UniqueName: \"kubernetes.io/projected/db35f8f2-75a6-4c65-a44b-09230c27f953-kube-api-access-7tf59\") pod \"db35f8f2-75a6-4c65-a44b-09230c27f953\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.472174 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-ssh-key\") pod \"db35f8f2-75a6-4c65-a44b-09230c27f953\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.472994 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-inventory\") pod \"db35f8f2-75a6-4c65-a44b-09230c27f953\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.473311 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-logging-compute-config-data-1\") pod \"db35f8f2-75a6-4c65-a44b-09230c27f953\" (UID: \"db35f8f2-75a6-4c65-a44b-09230c27f953\") " Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.493603 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db35f8f2-75a6-4c65-a44b-09230c27f953-kube-api-access-7tf59" (OuterVolumeSpecName: "kube-api-access-7tf59") pod "db35f8f2-75a6-4c65-a44b-09230c27f953" (UID: "db35f8f2-75a6-4c65-a44b-09230c27f953"). InnerVolumeSpecName "kube-api-access-7tf59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.508293 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "db35f8f2-75a6-4c65-a44b-09230c27f953" (UID: "db35f8f2-75a6-4c65-a44b-09230c27f953"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.510841 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "db35f8f2-75a6-4c65-a44b-09230c27f953" (UID: "db35f8f2-75a6-4c65-a44b-09230c27f953"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.528609 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "db35f8f2-75a6-4c65-a44b-09230c27f953" (UID: "db35f8f2-75a6-4c65-a44b-09230c27f953"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.539167 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-inventory" (OuterVolumeSpecName: "inventory") pod "db35f8f2-75a6-4c65-a44b-09230c27f953" (UID: "db35f8f2-75a6-4c65-a44b-09230c27f953"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.576559 4771 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.576597 4771 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.576611 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tf59\" (UniqueName: \"kubernetes.io/projected/db35f8f2-75a6-4c65-a44b-09230c27f953-kube-api-access-7tf59\") on node \"crc\" DevicePath \"\"" Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.576623 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.576642 4771 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db35f8f2-75a6-4c65-a44b-09230c27f953-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.848891 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" event={"ID":"db35f8f2-75a6-4c65-a44b-09230c27f953","Type":"ContainerDied","Data":"83bc8b240a061d77ab7f6bbe1935030cee17c42548e1a45f7d87ea7db2ff8257"} Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.848958 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83bc8b240a061d77ab7f6bbe1935030cee17c42548e1a45f7d87ea7db2ff8257" Oct 02 10:35:01 crc kubenswrapper[4771]: I1002 10:35:01.849014 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pczb2" Oct 02 10:36:41 crc kubenswrapper[4771]: E1002 10:36:41.176950 4771 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.53:50666->38.102.83.53:41045: write tcp 38.102.83.53:50666->38.102.83.53:41045: write: connection reset by peer Oct 02 10:36:42 crc kubenswrapper[4771]: I1002 10:36:42.151646 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:36:42 crc kubenswrapper[4771]: I1002 10:36:42.151978 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:37:12 crc kubenswrapper[4771]: I1002 10:37:12.145903 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:37:12 crc kubenswrapper[4771]: I1002 10:37:12.146512 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.207022 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x2jf9"] Oct 02 10:37:38 crc kubenswrapper[4771]: E1002 10:37:38.208271 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db35f8f2-75a6-4c65-a44b-09230c27f953" containerName="logging-edpm-deployment-openstack-edpm-ipam" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.208290 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="db35f8f2-75a6-4c65-a44b-09230c27f953" containerName="logging-edpm-deployment-openstack-edpm-ipam" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.208640 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="db35f8f2-75a6-4c65-a44b-09230c27f953" containerName="logging-edpm-deployment-openstack-edpm-ipam" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.211003 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.218981 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x2jf9"] Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.298110 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/525f1c5d-f5df-43f8-9d31-f116550c426a-utilities\") pod \"community-operators-x2jf9\" (UID: \"525f1c5d-f5df-43f8-9d31-f116550c426a\") " pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.298403 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/525f1c5d-f5df-43f8-9d31-f116550c426a-catalog-content\") pod \"community-operators-x2jf9\" (UID: \"525f1c5d-f5df-43f8-9d31-f116550c426a\") " pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.298774 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kz4p\" (UniqueName: \"kubernetes.io/projected/525f1c5d-f5df-43f8-9d31-f116550c426a-kube-api-access-5kz4p\") pod \"community-operators-x2jf9\" (UID: \"525f1c5d-f5df-43f8-9d31-f116550c426a\") " pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.401224 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/525f1c5d-f5df-43f8-9d31-f116550c426a-catalog-content\") pod \"community-operators-x2jf9\" (UID: \"525f1c5d-f5df-43f8-9d31-f116550c426a\") " pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.401361 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kz4p\" (UniqueName: \"kubernetes.io/projected/525f1c5d-f5df-43f8-9d31-f116550c426a-kube-api-access-5kz4p\") pod \"community-operators-x2jf9\" (UID: \"525f1c5d-f5df-43f8-9d31-f116550c426a\") " pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.401489 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/525f1c5d-f5df-43f8-9d31-f116550c426a-utilities\") pod \"community-operators-x2jf9\" (UID: \"525f1c5d-f5df-43f8-9d31-f116550c426a\") " pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.401964 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/525f1c5d-f5df-43f8-9d31-f116550c426a-utilities\") pod \"community-operators-x2jf9\" (UID: \"525f1c5d-f5df-43f8-9d31-f116550c426a\") " pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.402273 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/525f1c5d-f5df-43f8-9d31-f116550c426a-catalog-content\") pod \"community-operators-x2jf9\" (UID: \"525f1c5d-f5df-43f8-9d31-f116550c426a\") " pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.442551 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kz4p\" (UniqueName: \"kubernetes.io/projected/525f1c5d-f5df-43f8-9d31-f116550c426a-kube-api-access-5kz4p\") pod \"community-operators-x2jf9\" (UID: \"525f1c5d-f5df-43f8-9d31-f116550c426a\") " pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:38 crc kubenswrapper[4771]: I1002 10:37:38.534665 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:39 crc kubenswrapper[4771]: I1002 10:37:39.157445 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x2jf9"] Oct 02 10:37:39 crc kubenswrapper[4771]: I1002 10:37:39.777453 4771 generic.go:334] "Generic (PLEG): container finished" podID="525f1c5d-f5df-43f8-9d31-f116550c426a" containerID="b1fe90b44cf0d8920523eb71870ceac9315074c33f87dc443516d808e41add39" exitCode=0 Oct 02 10:37:39 crc kubenswrapper[4771]: I1002 10:37:39.777493 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x2jf9" event={"ID":"525f1c5d-f5df-43f8-9d31-f116550c426a","Type":"ContainerDied","Data":"b1fe90b44cf0d8920523eb71870ceac9315074c33f87dc443516d808e41add39"} Oct 02 10:37:39 crc kubenswrapper[4771]: I1002 10:37:39.778433 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x2jf9" event={"ID":"525f1c5d-f5df-43f8-9d31-f116550c426a","Type":"ContainerStarted","Data":"1eae9c4d1bfdab9c9022c51f7d4a8a00c47a3d3a259e6e0d0de4d610ccc9fefd"} Oct 02 10:37:42 crc kubenswrapper[4771]: I1002 10:37:42.145978 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:37:42 crc kubenswrapper[4771]: I1002 10:37:42.146666 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:37:42 crc kubenswrapper[4771]: I1002 10:37:42.146719 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 10:37:42 crc kubenswrapper[4771]: I1002 10:37:42.147584 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:37:42 crc kubenswrapper[4771]: I1002 10:37:42.147638 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" gracePeriod=600 Oct 02 10:37:42 crc kubenswrapper[4771]: E1002 10:37:42.348631 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:37:42 crc kubenswrapper[4771]: I1002 10:37:42.811768 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" exitCode=0 Oct 02 10:37:42 crc kubenswrapper[4771]: I1002 10:37:42.811810 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf"} Oct 02 10:37:42 crc kubenswrapper[4771]: I1002 10:37:42.811850 4771 scope.go:117] "RemoveContainer" containerID="38f585bfa8e4f2dadd77f4304ba67b1b2f4b224957e4927913f6b7f916a67b45" Oct 02 10:37:42 crc kubenswrapper[4771]: I1002 10:37:42.812661 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:37:42 crc kubenswrapper[4771]: E1002 10:37:42.813074 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:37:43 crc kubenswrapper[4771]: I1002 10:37:43.823019 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x2jf9" event={"ID":"525f1c5d-f5df-43f8-9d31-f116550c426a","Type":"ContainerStarted","Data":"b163c0ed9c352d47a0ba71eaf41914dd488d9c6e2ff730cb6ec4530f5f8e358d"} Oct 02 10:37:47 crc kubenswrapper[4771]: I1002 10:37:47.878936 4771 generic.go:334] "Generic (PLEG): container finished" podID="525f1c5d-f5df-43f8-9d31-f116550c426a" containerID="b163c0ed9c352d47a0ba71eaf41914dd488d9c6e2ff730cb6ec4530f5f8e358d" exitCode=0 Oct 02 10:37:47 crc kubenswrapper[4771]: I1002 10:37:47.879030 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x2jf9" event={"ID":"525f1c5d-f5df-43f8-9d31-f116550c426a","Type":"ContainerDied","Data":"b163c0ed9c352d47a0ba71eaf41914dd488d9c6e2ff730cb6ec4530f5f8e358d"} Oct 02 10:37:48 crc kubenswrapper[4771]: I1002 10:37:48.894204 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x2jf9" event={"ID":"525f1c5d-f5df-43f8-9d31-f116550c426a","Type":"ContainerStarted","Data":"ce50527573af96040415664b094393c2667e0e2141b6779b3b2c4b0437873929"} Oct 02 10:37:48 crc kubenswrapper[4771]: I1002 10:37:48.920460 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x2jf9" podStartSLOduration=2.224941225 podStartE2EDuration="10.920439003s" podCreationTimestamp="2025-10-02 10:37:38 +0000 UTC" firstStartedPulling="2025-10-02 10:37:39.779526309 +0000 UTC m=+3647.427211376" lastFinishedPulling="2025-10-02 10:37:48.475024087 +0000 UTC m=+3656.122709154" observedRunningTime="2025-10-02 10:37:48.919545311 +0000 UTC m=+3656.567230388" watchObservedRunningTime="2025-10-02 10:37:48.920439003 +0000 UTC m=+3656.568124070" Oct 02 10:37:54 crc kubenswrapper[4771]: I1002 10:37:54.682691 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:37:54 crc kubenswrapper[4771]: E1002 10:37:54.684104 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:37:58 crc kubenswrapper[4771]: I1002 10:37:58.535317 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:58 crc kubenswrapper[4771]: I1002 10:37:58.535640 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:58 crc kubenswrapper[4771]: I1002 10:37:58.589750 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:59 crc kubenswrapper[4771]: I1002 10:37:59.122770 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:37:59 crc kubenswrapper[4771]: I1002 10:37:59.169664 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x2jf9"] Oct 02 10:38:01 crc kubenswrapper[4771]: I1002 10:38:01.035726 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x2jf9" podUID="525f1c5d-f5df-43f8-9d31-f116550c426a" containerName="registry-server" containerID="cri-o://ce50527573af96040415664b094393c2667e0e2141b6779b3b2c4b0437873929" gracePeriod=2 Oct 02 10:38:01 crc kubenswrapper[4771]: I1002 10:38:01.683667 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:38:01 crc kubenswrapper[4771]: I1002 10:38:01.882034 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/525f1c5d-f5df-43f8-9d31-f116550c426a-catalog-content\") pod \"525f1c5d-f5df-43f8-9d31-f116550c426a\" (UID: \"525f1c5d-f5df-43f8-9d31-f116550c426a\") " Oct 02 10:38:01 crc kubenswrapper[4771]: I1002 10:38:01.882565 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kz4p\" (UniqueName: \"kubernetes.io/projected/525f1c5d-f5df-43f8-9d31-f116550c426a-kube-api-access-5kz4p\") pod \"525f1c5d-f5df-43f8-9d31-f116550c426a\" (UID: \"525f1c5d-f5df-43f8-9d31-f116550c426a\") " Oct 02 10:38:01 crc kubenswrapper[4771]: I1002 10:38:01.882604 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/525f1c5d-f5df-43f8-9d31-f116550c426a-utilities\") pod \"525f1c5d-f5df-43f8-9d31-f116550c426a\" (UID: \"525f1c5d-f5df-43f8-9d31-f116550c426a\") " Oct 02 10:38:01 crc kubenswrapper[4771]: I1002 10:38:01.883358 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/525f1c5d-f5df-43f8-9d31-f116550c426a-utilities" (OuterVolumeSpecName: "utilities") pod "525f1c5d-f5df-43f8-9d31-f116550c426a" (UID: "525f1c5d-f5df-43f8-9d31-f116550c426a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:38:01 crc kubenswrapper[4771]: I1002 10:38:01.887987 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/525f1c5d-f5df-43f8-9d31-f116550c426a-kube-api-access-5kz4p" (OuterVolumeSpecName: "kube-api-access-5kz4p") pod "525f1c5d-f5df-43f8-9d31-f116550c426a" (UID: "525f1c5d-f5df-43f8-9d31-f116550c426a"). InnerVolumeSpecName "kube-api-access-5kz4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:38:01 crc kubenswrapper[4771]: I1002 10:38:01.933712 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/525f1c5d-f5df-43f8-9d31-f116550c426a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "525f1c5d-f5df-43f8-9d31-f116550c426a" (UID: "525f1c5d-f5df-43f8-9d31-f116550c426a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:38:01 crc kubenswrapper[4771]: I1002 10:38:01.984992 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kz4p\" (UniqueName: \"kubernetes.io/projected/525f1c5d-f5df-43f8-9d31-f116550c426a-kube-api-access-5kz4p\") on node \"crc\" DevicePath \"\"" Oct 02 10:38:01 crc kubenswrapper[4771]: I1002 10:38:01.985029 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/525f1c5d-f5df-43f8-9d31-f116550c426a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:38:01 crc kubenswrapper[4771]: I1002 10:38:01.985038 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/525f1c5d-f5df-43f8-9d31-f116550c426a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.051262 4771 generic.go:334] "Generic (PLEG): container finished" podID="525f1c5d-f5df-43f8-9d31-f116550c426a" containerID="ce50527573af96040415664b094393c2667e0e2141b6779b3b2c4b0437873929" exitCode=0 Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.051310 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x2jf9" event={"ID":"525f1c5d-f5df-43f8-9d31-f116550c426a","Type":"ContainerDied","Data":"ce50527573af96040415664b094393c2667e0e2141b6779b3b2c4b0437873929"} Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.051330 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x2jf9" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.051345 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x2jf9" event={"ID":"525f1c5d-f5df-43f8-9d31-f116550c426a","Type":"ContainerDied","Data":"1eae9c4d1bfdab9c9022c51f7d4a8a00c47a3d3a259e6e0d0de4d610ccc9fefd"} Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.051368 4771 scope.go:117] "RemoveContainer" containerID="ce50527573af96040415664b094393c2667e0e2141b6779b3b2c4b0437873929" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.088515 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x2jf9"] Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.098987 4771 scope.go:117] "RemoveContainer" containerID="b163c0ed9c352d47a0ba71eaf41914dd488d9c6e2ff730cb6ec4530f5f8e358d" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.105749 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x2jf9"] Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.127781 4771 scope.go:117] "RemoveContainer" containerID="b1fe90b44cf0d8920523eb71870ceac9315074c33f87dc443516d808e41add39" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.202060 4771 scope.go:117] "RemoveContainer" containerID="ce50527573af96040415664b094393c2667e0e2141b6779b3b2c4b0437873929" Oct 02 10:38:02 crc kubenswrapper[4771]: E1002 10:38:02.203621 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce50527573af96040415664b094393c2667e0e2141b6779b3b2c4b0437873929\": container with ID starting with ce50527573af96040415664b094393c2667e0e2141b6779b3b2c4b0437873929 not found: ID does not exist" containerID="ce50527573af96040415664b094393c2667e0e2141b6779b3b2c4b0437873929" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.203666 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce50527573af96040415664b094393c2667e0e2141b6779b3b2c4b0437873929"} err="failed to get container status \"ce50527573af96040415664b094393c2667e0e2141b6779b3b2c4b0437873929\": rpc error: code = NotFound desc = could not find container \"ce50527573af96040415664b094393c2667e0e2141b6779b3b2c4b0437873929\": container with ID starting with ce50527573af96040415664b094393c2667e0e2141b6779b3b2c4b0437873929 not found: ID does not exist" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.203693 4771 scope.go:117] "RemoveContainer" containerID="b163c0ed9c352d47a0ba71eaf41914dd488d9c6e2ff730cb6ec4530f5f8e358d" Oct 02 10:38:02 crc kubenswrapper[4771]: E1002 10:38:02.204109 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b163c0ed9c352d47a0ba71eaf41914dd488d9c6e2ff730cb6ec4530f5f8e358d\": container with ID starting with b163c0ed9c352d47a0ba71eaf41914dd488d9c6e2ff730cb6ec4530f5f8e358d not found: ID does not exist" containerID="b163c0ed9c352d47a0ba71eaf41914dd488d9c6e2ff730cb6ec4530f5f8e358d" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.204174 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b163c0ed9c352d47a0ba71eaf41914dd488d9c6e2ff730cb6ec4530f5f8e358d"} err="failed to get container status \"b163c0ed9c352d47a0ba71eaf41914dd488d9c6e2ff730cb6ec4530f5f8e358d\": rpc error: code = NotFound desc = could not find container \"b163c0ed9c352d47a0ba71eaf41914dd488d9c6e2ff730cb6ec4530f5f8e358d\": container with ID starting with b163c0ed9c352d47a0ba71eaf41914dd488d9c6e2ff730cb6ec4530f5f8e358d not found: ID does not exist" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.204189 4771 scope.go:117] "RemoveContainer" containerID="b1fe90b44cf0d8920523eb71870ceac9315074c33f87dc443516d808e41add39" Oct 02 10:38:02 crc kubenswrapper[4771]: E1002 10:38:02.204596 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1fe90b44cf0d8920523eb71870ceac9315074c33f87dc443516d808e41add39\": container with ID starting with b1fe90b44cf0d8920523eb71870ceac9315074c33f87dc443516d808e41add39 not found: ID does not exist" containerID="b1fe90b44cf0d8920523eb71870ceac9315074c33f87dc443516d808e41add39" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.204631 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1fe90b44cf0d8920523eb71870ceac9315074c33f87dc443516d808e41add39"} err="failed to get container status \"b1fe90b44cf0d8920523eb71870ceac9315074c33f87dc443516d808e41add39\": rpc error: code = NotFound desc = could not find container \"b1fe90b44cf0d8920523eb71870ceac9315074c33f87dc443516d808e41add39\": container with ID starting with b1fe90b44cf0d8920523eb71870ceac9315074c33f87dc443516d808e41add39 not found: ID does not exist" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.243992 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v44nr"] Oct 02 10:38:02 crc kubenswrapper[4771]: E1002 10:38:02.244477 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="525f1c5d-f5df-43f8-9d31-f116550c426a" containerName="extract-content" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.244516 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="525f1c5d-f5df-43f8-9d31-f116550c426a" containerName="extract-content" Oct 02 10:38:02 crc kubenswrapper[4771]: E1002 10:38:02.244540 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="525f1c5d-f5df-43f8-9d31-f116550c426a" containerName="registry-server" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.244547 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="525f1c5d-f5df-43f8-9d31-f116550c426a" containerName="registry-server" Oct 02 10:38:02 crc kubenswrapper[4771]: E1002 10:38:02.244581 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="525f1c5d-f5df-43f8-9d31-f116550c426a" containerName="extract-utilities" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.244589 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="525f1c5d-f5df-43f8-9d31-f116550c426a" containerName="extract-utilities" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.244826 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="525f1c5d-f5df-43f8-9d31-f116550c426a" containerName="registry-server" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.249210 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.259024 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v44nr"] Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.294281 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brnwn\" (UniqueName: \"kubernetes.io/projected/96ce116f-5bf5-4b31-90df-677371908ad5-kube-api-access-brnwn\") pod \"redhat-marketplace-v44nr\" (UID: \"96ce116f-5bf5-4b31-90df-677371908ad5\") " pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.294726 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96ce116f-5bf5-4b31-90df-677371908ad5-catalog-content\") pod \"redhat-marketplace-v44nr\" (UID: \"96ce116f-5bf5-4b31-90df-677371908ad5\") " pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.294836 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96ce116f-5bf5-4b31-90df-677371908ad5-utilities\") pod \"redhat-marketplace-v44nr\" (UID: \"96ce116f-5bf5-4b31-90df-677371908ad5\") " pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.396448 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96ce116f-5bf5-4b31-90df-677371908ad5-catalog-content\") pod \"redhat-marketplace-v44nr\" (UID: \"96ce116f-5bf5-4b31-90df-677371908ad5\") " pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.396735 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96ce116f-5bf5-4b31-90df-677371908ad5-utilities\") pod \"redhat-marketplace-v44nr\" (UID: \"96ce116f-5bf5-4b31-90df-677371908ad5\") " pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.396901 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brnwn\" (UniqueName: \"kubernetes.io/projected/96ce116f-5bf5-4b31-90df-677371908ad5-kube-api-access-brnwn\") pod \"redhat-marketplace-v44nr\" (UID: \"96ce116f-5bf5-4b31-90df-677371908ad5\") " pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.396920 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96ce116f-5bf5-4b31-90df-677371908ad5-catalog-content\") pod \"redhat-marketplace-v44nr\" (UID: \"96ce116f-5bf5-4b31-90df-677371908ad5\") " pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.397121 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96ce116f-5bf5-4b31-90df-677371908ad5-utilities\") pod \"redhat-marketplace-v44nr\" (UID: \"96ce116f-5bf5-4b31-90df-677371908ad5\") " pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.423204 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brnwn\" (UniqueName: \"kubernetes.io/projected/96ce116f-5bf5-4b31-90df-677371908ad5-kube-api-access-brnwn\") pod \"redhat-marketplace-v44nr\" (UID: \"96ce116f-5bf5-4b31-90df-677371908ad5\") " pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:02 crc kubenswrapper[4771]: I1002 10:38:02.584494 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:03 crc kubenswrapper[4771]: I1002 10:38:03.119729 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v44nr"] Oct 02 10:38:03 crc kubenswrapper[4771]: I1002 10:38:03.698919 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="525f1c5d-f5df-43f8-9d31-f116550c426a" path="/var/lib/kubelet/pods/525f1c5d-f5df-43f8-9d31-f116550c426a/volumes" Oct 02 10:38:04 crc kubenswrapper[4771]: I1002 10:38:04.076920 4771 generic.go:334] "Generic (PLEG): container finished" podID="96ce116f-5bf5-4b31-90df-677371908ad5" containerID="f099b578b158c2fb8773bfb88d4fbd0b04e20e1454bcd092cbe72c35bbb49400" exitCode=0 Oct 02 10:38:04 crc kubenswrapper[4771]: I1002 10:38:04.076970 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v44nr" event={"ID":"96ce116f-5bf5-4b31-90df-677371908ad5","Type":"ContainerDied","Data":"f099b578b158c2fb8773bfb88d4fbd0b04e20e1454bcd092cbe72c35bbb49400"} Oct 02 10:38:04 crc kubenswrapper[4771]: I1002 10:38:04.077003 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v44nr" event={"ID":"96ce116f-5bf5-4b31-90df-677371908ad5","Type":"ContainerStarted","Data":"a09f9036426330693509b8fd174731cfc24a9956e811d73acc02bc94369135de"} Oct 02 10:38:06 crc kubenswrapper[4771]: I1002 10:38:06.098924 4771 generic.go:334] "Generic (PLEG): container finished" podID="96ce116f-5bf5-4b31-90df-677371908ad5" containerID="1ab732af8a72d4529c92f55757b77ff6447680f1a49ff121dd99b1594d97174c" exitCode=0 Oct 02 10:38:06 crc kubenswrapper[4771]: I1002 10:38:06.099451 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v44nr" event={"ID":"96ce116f-5bf5-4b31-90df-677371908ad5","Type":"ContainerDied","Data":"1ab732af8a72d4529c92f55757b77ff6447680f1a49ff121dd99b1594d97174c"} Oct 02 10:38:07 crc kubenswrapper[4771]: I1002 10:38:07.113607 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v44nr" event={"ID":"96ce116f-5bf5-4b31-90df-677371908ad5","Type":"ContainerStarted","Data":"1378bc706c7b79911f17734515f36b7aace162810004fe4301002f3f2b9ad87e"} Oct 02 10:38:07 crc kubenswrapper[4771]: I1002 10:38:07.140070 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v44nr" podStartSLOduration=2.582679732 podStartE2EDuration="5.14004799s" podCreationTimestamp="2025-10-02 10:38:02 +0000 UTC" firstStartedPulling="2025-10-02 10:38:04.079325782 +0000 UTC m=+3671.727010849" lastFinishedPulling="2025-10-02 10:38:06.63669405 +0000 UTC m=+3674.284379107" observedRunningTime="2025-10-02 10:38:07.130957781 +0000 UTC m=+3674.778642848" watchObservedRunningTime="2025-10-02 10:38:07.14004799 +0000 UTC m=+3674.787733057" Oct 02 10:38:08 crc kubenswrapper[4771]: I1002 10:38:08.682947 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:38:08 crc kubenswrapper[4771]: E1002 10:38:08.683557 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:38:12 crc kubenswrapper[4771]: I1002 10:38:12.586176 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:12 crc kubenswrapper[4771]: I1002 10:38:12.586702 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:12 crc kubenswrapper[4771]: I1002 10:38:12.638258 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:13 crc kubenswrapper[4771]: I1002 10:38:13.236318 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:13 crc kubenswrapper[4771]: I1002 10:38:13.290471 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v44nr"] Oct 02 10:38:15 crc kubenswrapper[4771]: I1002 10:38:15.196898 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v44nr" podUID="96ce116f-5bf5-4b31-90df-677371908ad5" containerName="registry-server" containerID="cri-o://1378bc706c7b79911f17734515f36b7aace162810004fe4301002f3f2b9ad87e" gracePeriod=2 Oct 02 10:38:15 crc kubenswrapper[4771]: I1002 10:38:15.739803 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:15 crc kubenswrapper[4771]: I1002 10:38:15.838936 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96ce116f-5bf5-4b31-90df-677371908ad5-catalog-content\") pod \"96ce116f-5bf5-4b31-90df-677371908ad5\" (UID: \"96ce116f-5bf5-4b31-90df-677371908ad5\") " Oct 02 10:38:15 crc kubenswrapper[4771]: I1002 10:38:15.839112 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brnwn\" (UniqueName: \"kubernetes.io/projected/96ce116f-5bf5-4b31-90df-677371908ad5-kube-api-access-brnwn\") pod \"96ce116f-5bf5-4b31-90df-677371908ad5\" (UID: \"96ce116f-5bf5-4b31-90df-677371908ad5\") " Oct 02 10:38:15 crc kubenswrapper[4771]: I1002 10:38:15.839269 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96ce116f-5bf5-4b31-90df-677371908ad5-utilities\") pod \"96ce116f-5bf5-4b31-90df-677371908ad5\" (UID: \"96ce116f-5bf5-4b31-90df-677371908ad5\") " Oct 02 10:38:15 crc kubenswrapper[4771]: I1002 10:38:15.839996 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96ce116f-5bf5-4b31-90df-677371908ad5-utilities" (OuterVolumeSpecName: "utilities") pod "96ce116f-5bf5-4b31-90df-677371908ad5" (UID: "96ce116f-5bf5-4b31-90df-677371908ad5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:38:15 crc kubenswrapper[4771]: I1002 10:38:15.847864 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96ce116f-5bf5-4b31-90df-677371908ad5-kube-api-access-brnwn" (OuterVolumeSpecName: "kube-api-access-brnwn") pod "96ce116f-5bf5-4b31-90df-677371908ad5" (UID: "96ce116f-5bf5-4b31-90df-677371908ad5"). InnerVolumeSpecName "kube-api-access-brnwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:38:15 crc kubenswrapper[4771]: I1002 10:38:15.857510 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96ce116f-5bf5-4b31-90df-677371908ad5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96ce116f-5bf5-4b31-90df-677371908ad5" (UID: "96ce116f-5bf5-4b31-90df-677371908ad5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:38:15 crc kubenswrapper[4771]: I1002 10:38:15.942780 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96ce116f-5bf5-4b31-90df-677371908ad5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:38:15 crc kubenswrapper[4771]: I1002 10:38:15.942829 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brnwn\" (UniqueName: \"kubernetes.io/projected/96ce116f-5bf5-4b31-90df-677371908ad5-kube-api-access-brnwn\") on node \"crc\" DevicePath \"\"" Oct 02 10:38:15 crc kubenswrapper[4771]: I1002 10:38:15.942840 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96ce116f-5bf5-4b31-90df-677371908ad5-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.209040 4771 generic.go:334] "Generic (PLEG): container finished" podID="96ce116f-5bf5-4b31-90df-677371908ad5" containerID="1378bc706c7b79911f17734515f36b7aace162810004fe4301002f3f2b9ad87e" exitCode=0 Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.209092 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v44nr" event={"ID":"96ce116f-5bf5-4b31-90df-677371908ad5","Type":"ContainerDied","Data":"1378bc706c7b79911f17734515f36b7aace162810004fe4301002f3f2b9ad87e"} Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.209119 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v44nr" Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.209144 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v44nr" event={"ID":"96ce116f-5bf5-4b31-90df-677371908ad5","Type":"ContainerDied","Data":"a09f9036426330693509b8fd174731cfc24a9956e811d73acc02bc94369135de"} Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.209180 4771 scope.go:117] "RemoveContainer" containerID="1378bc706c7b79911f17734515f36b7aace162810004fe4301002f3f2b9ad87e" Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.243385 4771 scope.go:117] "RemoveContainer" containerID="1ab732af8a72d4529c92f55757b77ff6447680f1a49ff121dd99b1594d97174c" Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.258184 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v44nr"] Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.272523 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v44nr"] Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.278576 4771 scope.go:117] "RemoveContainer" containerID="f099b578b158c2fb8773bfb88d4fbd0b04e20e1454bcd092cbe72c35bbb49400" Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.335476 4771 scope.go:117] "RemoveContainer" containerID="1378bc706c7b79911f17734515f36b7aace162810004fe4301002f3f2b9ad87e" Oct 02 10:38:16 crc kubenswrapper[4771]: E1002 10:38:16.335820 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1378bc706c7b79911f17734515f36b7aace162810004fe4301002f3f2b9ad87e\": container with ID starting with 1378bc706c7b79911f17734515f36b7aace162810004fe4301002f3f2b9ad87e not found: ID does not exist" containerID="1378bc706c7b79911f17734515f36b7aace162810004fe4301002f3f2b9ad87e" Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.335859 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1378bc706c7b79911f17734515f36b7aace162810004fe4301002f3f2b9ad87e"} err="failed to get container status \"1378bc706c7b79911f17734515f36b7aace162810004fe4301002f3f2b9ad87e\": rpc error: code = NotFound desc = could not find container \"1378bc706c7b79911f17734515f36b7aace162810004fe4301002f3f2b9ad87e\": container with ID starting with 1378bc706c7b79911f17734515f36b7aace162810004fe4301002f3f2b9ad87e not found: ID does not exist" Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.335883 4771 scope.go:117] "RemoveContainer" containerID="1ab732af8a72d4529c92f55757b77ff6447680f1a49ff121dd99b1594d97174c" Oct 02 10:38:16 crc kubenswrapper[4771]: E1002 10:38:16.336309 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ab732af8a72d4529c92f55757b77ff6447680f1a49ff121dd99b1594d97174c\": container with ID starting with 1ab732af8a72d4529c92f55757b77ff6447680f1a49ff121dd99b1594d97174c not found: ID does not exist" containerID="1ab732af8a72d4529c92f55757b77ff6447680f1a49ff121dd99b1594d97174c" Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.336354 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ab732af8a72d4529c92f55757b77ff6447680f1a49ff121dd99b1594d97174c"} err="failed to get container status \"1ab732af8a72d4529c92f55757b77ff6447680f1a49ff121dd99b1594d97174c\": rpc error: code = NotFound desc = could not find container \"1ab732af8a72d4529c92f55757b77ff6447680f1a49ff121dd99b1594d97174c\": container with ID starting with 1ab732af8a72d4529c92f55757b77ff6447680f1a49ff121dd99b1594d97174c not found: ID does not exist" Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.336399 4771 scope.go:117] "RemoveContainer" containerID="f099b578b158c2fb8773bfb88d4fbd0b04e20e1454bcd092cbe72c35bbb49400" Oct 02 10:38:16 crc kubenswrapper[4771]: E1002 10:38:16.336635 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f099b578b158c2fb8773bfb88d4fbd0b04e20e1454bcd092cbe72c35bbb49400\": container with ID starting with f099b578b158c2fb8773bfb88d4fbd0b04e20e1454bcd092cbe72c35bbb49400 not found: ID does not exist" containerID="f099b578b158c2fb8773bfb88d4fbd0b04e20e1454bcd092cbe72c35bbb49400" Oct 02 10:38:16 crc kubenswrapper[4771]: I1002 10:38:16.336656 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f099b578b158c2fb8773bfb88d4fbd0b04e20e1454bcd092cbe72c35bbb49400"} err="failed to get container status \"f099b578b158c2fb8773bfb88d4fbd0b04e20e1454bcd092cbe72c35bbb49400\": rpc error: code = NotFound desc = could not find container \"f099b578b158c2fb8773bfb88d4fbd0b04e20e1454bcd092cbe72c35bbb49400\": container with ID starting with f099b578b158c2fb8773bfb88d4fbd0b04e20e1454bcd092cbe72c35bbb49400 not found: ID does not exist" Oct 02 10:38:17 crc kubenswrapper[4771]: I1002 10:38:17.705867 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96ce116f-5bf5-4b31-90df-677371908ad5" path="/var/lib/kubelet/pods/96ce116f-5bf5-4b31-90df-677371908ad5/volumes" Oct 02 10:38:20 crc kubenswrapper[4771]: I1002 10:38:20.681883 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:38:20 crc kubenswrapper[4771]: E1002 10:38:20.682655 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:38:31 crc kubenswrapper[4771]: I1002 10:38:31.681500 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:38:31 crc kubenswrapper[4771]: E1002 10:38:31.682284 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:38:45 crc kubenswrapper[4771]: I1002 10:38:45.682209 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:38:45 crc kubenswrapper[4771]: E1002 10:38:45.683400 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:38:59 crc kubenswrapper[4771]: I1002 10:38:59.681693 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:38:59 crc kubenswrapper[4771]: E1002 10:38:59.682528 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:39:12 crc kubenswrapper[4771]: I1002 10:39:12.682295 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:39:12 crc kubenswrapper[4771]: E1002 10:39:12.683462 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:39:25 crc kubenswrapper[4771]: I1002 10:39:25.681251 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:39:25 crc kubenswrapper[4771]: E1002 10:39:25.682034 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:39:40 crc kubenswrapper[4771]: I1002 10:39:40.681851 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:39:40 crc kubenswrapper[4771]: E1002 10:39:40.683161 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:39:54 crc kubenswrapper[4771]: I1002 10:39:54.681801 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:39:54 crc kubenswrapper[4771]: E1002 10:39:54.682647 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:40:08 crc kubenswrapper[4771]: I1002 10:40:08.682593 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:40:08 crc kubenswrapper[4771]: E1002 10:40:08.683911 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:40:21 crc kubenswrapper[4771]: I1002 10:40:21.687894 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:40:21 crc kubenswrapper[4771]: E1002 10:40:21.688825 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:40:33 crc kubenswrapper[4771]: I1002 10:40:33.693345 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:40:33 crc kubenswrapper[4771]: E1002 10:40:33.694117 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:40:34 crc kubenswrapper[4771]: E1002 10:40:34.419205 4771 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.53:55176->38.102.83.53:41045: read tcp 38.102.83.53:55176->38.102.83.53:41045: read: connection reset by peer Oct 02 10:40:34 crc kubenswrapper[4771]: E1002 10:40:34.419582 4771 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.53:55176->38.102.83.53:41045: write tcp 38.102.83.53:55176->38.102.83.53:41045: write: broken pipe Oct 02 10:40:46 crc kubenswrapper[4771]: I1002 10:40:46.681711 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:40:46 crc kubenswrapper[4771]: E1002 10:40:46.683088 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:41:00 crc kubenswrapper[4771]: I1002 10:41:00.681337 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:41:00 crc kubenswrapper[4771]: E1002 10:41:00.682080 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:41:14 crc kubenswrapper[4771]: I1002 10:41:14.682744 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:41:14 crc kubenswrapper[4771]: E1002 10:41:14.684653 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:41:25 crc kubenswrapper[4771]: I1002 10:41:25.681895 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:41:25 crc kubenswrapper[4771]: E1002 10:41:25.682838 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:41:38 crc kubenswrapper[4771]: I1002 10:41:38.681270 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:41:38 crc kubenswrapper[4771]: E1002 10:41:38.682472 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:41:49 crc kubenswrapper[4771]: I1002 10:41:49.683971 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:41:49 crc kubenswrapper[4771]: E1002 10:41:49.687387 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:42:00 crc kubenswrapper[4771]: I1002 10:42:00.681970 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:42:00 crc kubenswrapper[4771]: E1002 10:42:00.683060 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:42:15 crc kubenswrapper[4771]: I1002 10:42:15.681356 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:42:15 crc kubenswrapper[4771]: E1002 10:42:15.682258 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:42:26 crc kubenswrapper[4771]: I1002 10:42:26.682025 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:42:26 crc kubenswrapper[4771]: E1002 10:42:26.682932 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:42:38 crc kubenswrapper[4771]: I1002 10:42:38.683063 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:42:38 crc kubenswrapper[4771]: E1002 10:42:38.683859 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:42:53 crc kubenswrapper[4771]: I1002 10:42:53.696248 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:42:54 crc kubenswrapper[4771]: I1002 10:42:54.618290 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"194c5b729f6dfb0975af45a7ef28aedac3730958132074ff9a427ab736eddfcb"} Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.584509 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-67xqh"] Oct 02 10:43:29 crc kubenswrapper[4771]: E1002 10:43:29.585833 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96ce116f-5bf5-4b31-90df-677371908ad5" containerName="registry-server" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.585860 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="96ce116f-5bf5-4b31-90df-677371908ad5" containerName="registry-server" Oct 02 10:43:29 crc kubenswrapper[4771]: E1002 10:43:29.585885 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96ce116f-5bf5-4b31-90df-677371908ad5" containerName="extract-utilities" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.585894 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="96ce116f-5bf5-4b31-90df-677371908ad5" containerName="extract-utilities" Oct 02 10:43:29 crc kubenswrapper[4771]: E1002 10:43:29.585928 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96ce116f-5bf5-4b31-90df-677371908ad5" containerName="extract-content" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.585935 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="96ce116f-5bf5-4b31-90df-677371908ad5" containerName="extract-content" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.586259 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="96ce116f-5bf5-4b31-90df-677371908ad5" containerName="registry-server" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.589248 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.600314 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-67xqh"] Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.765323 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbrzr\" (UniqueName: \"kubernetes.io/projected/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-kube-api-access-kbrzr\") pod \"redhat-operators-67xqh\" (UID: \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\") " pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.765444 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-utilities\") pod \"redhat-operators-67xqh\" (UID: \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\") " pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.765628 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-catalog-content\") pod \"redhat-operators-67xqh\" (UID: \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\") " pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.868557 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-utilities\") pod \"redhat-operators-67xqh\" (UID: \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\") " pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.868937 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-catalog-content\") pod \"redhat-operators-67xqh\" (UID: \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\") " pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.869311 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-utilities\") pod \"redhat-operators-67xqh\" (UID: \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\") " pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.869348 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-catalog-content\") pod \"redhat-operators-67xqh\" (UID: \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\") " pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.869322 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbrzr\" (UniqueName: \"kubernetes.io/projected/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-kube-api-access-kbrzr\") pod \"redhat-operators-67xqh\" (UID: \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\") " pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.896303 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbrzr\" (UniqueName: \"kubernetes.io/projected/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-kube-api-access-kbrzr\") pod \"redhat-operators-67xqh\" (UID: \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\") " pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:29 crc kubenswrapper[4771]: I1002 10:43:29.922845 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:30 crc kubenswrapper[4771]: I1002 10:43:30.543982 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-67xqh"] Oct 02 10:43:31 crc kubenswrapper[4771]: I1002 10:43:31.113375 4771 generic.go:334] "Generic (PLEG): container finished" podID="1e820e3b-7b37-45ec-a57d-ebb07978d6fa" containerID="dadad7faca2b248d2535ad7034b000a1126aa48c8784fea2ca273ddcee7890c8" exitCode=0 Oct 02 10:43:31 crc kubenswrapper[4771]: I1002 10:43:31.116206 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67xqh" event={"ID":"1e820e3b-7b37-45ec-a57d-ebb07978d6fa","Type":"ContainerDied","Data":"dadad7faca2b248d2535ad7034b000a1126aa48c8784fea2ca273ddcee7890c8"} Oct 02 10:43:31 crc kubenswrapper[4771]: I1002 10:43:31.116280 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67xqh" event={"ID":"1e820e3b-7b37-45ec-a57d-ebb07978d6fa","Type":"ContainerStarted","Data":"1aaca6614d6d6a89a5b9b35db40e5fda3947586306bd9392431861962bb6cd83"} Oct 02 10:43:31 crc kubenswrapper[4771]: I1002 10:43:31.120911 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:43:32 crc kubenswrapper[4771]: I1002 10:43:32.144810 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67xqh" event={"ID":"1e820e3b-7b37-45ec-a57d-ebb07978d6fa","Type":"ContainerStarted","Data":"126e26a50268b991e0d0c29ec5107b7517b4e5f51672ccb015e3be8525548203"} Oct 02 10:43:33 crc kubenswrapper[4771]: I1002 10:43:33.173401 4771 generic.go:334] "Generic (PLEG): container finished" podID="1e820e3b-7b37-45ec-a57d-ebb07978d6fa" containerID="126e26a50268b991e0d0c29ec5107b7517b4e5f51672ccb015e3be8525548203" exitCode=0 Oct 02 10:43:33 crc kubenswrapper[4771]: I1002 10:43:33.173468 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67xqh" event={"ID":"1e820e3b-7b37-45ec-a57d-ebb07978d6fa","Type":"ContainerDied","Data":"126e26a50268b991e0d0c29ec5107b7517b4e5f51672ccb015e3be8525548203"} Oct 02 10:43:38 crc kubenswrapper[4771]: I1002 10:43:38.229425 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67xqh" event={"ID":"1e820e3b-7b37-45ec-a57d-ebb07978d6fa","Type":"ContainerStarted","Data":"0d63137206c8adab41354d0d591848fdf0211b5776ec2bb06faf62fb59ed875a"} Oct 02 10:43:39 crc kubenswrapper[4771]: I1002 10:43:39.923320 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:39 crc kubenswrapper[4771]: I1002 10:43:39.924088 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:40 crc kubenswrapper[4771]: I1002 10:43:40.980865 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-67xqh" podUID="1e820e3b-7b37-45ec-a57d-ebb07978d6fa" containerName="registry-server" probeResult="failure" output=< Oct 02 10:43:40 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:43:40 crc kubenswrapper[4771]: > Oct 02 10:43:49 crc kubenswrapper[4771]: I1002 10:43:49.986089 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:50 crc kubenswrapper[4771]: I1002 10:43:50.008255 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-67xqh" podStartSLOduration=14.801046278 podStartE2EDuration="21.00823238s" podCreationTimestamp="2025-10-02 10:43:29 +0000 UTC" firstStartedPulling="2025-10-02 10:43:31.119894094 +0000 UTC m=+3998.767579171" lastFinishedPulling="2025-10-02 10:43:37.327080206 +0000 UTC m=+4004.974765273" observedRunningTime="2025-10-02 10:43:38.254585412 +0000 UTC m=+4005.902270509" watchObservedRunningTime="2025-10-02 10:43:50.00823238 +0000 UTC m=+4017.655917447" Oct 02 10:43:50 crc kubenswrapper[4771]: I1002 10:43:50.042179 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:50 crc kubenswrapper[4771]: I1002 10:43:50.225425 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-67xqh"] Oct 02 10:43:51 crc kubenswrapper[4771]: I1002 10:43:51.401017 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-67xqh" podUID="1e820e3b-7b37-45ec-a57d-ebb07978d6fa" containerName="registry-server" containerID="cri-o://0d63137206c8adab41354d0d591848fdf0211b5776ec2bb06faf62fb59ed875a" gracePeriod=2 Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.066640 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.159229 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbrzr\" (UniqueName: \"kubernetes.io/projected/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-kube-api-access-kbrzr\") pod \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\" (UID: \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\") " Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.159478 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-utilities\") pod \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\" (UID: \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\") " Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.159535 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-catalog-content\") pod \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\" (UID: \"1e820e3b-7b37-45ec-a57d-ebb07978d6fa\") " Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.160478 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-utilities" (OuterVolumeSpecName: "utilities") pod "1e820e3b-7b37-45ec-a57d-ebb07978d6fa" (UID: "1e820e3b-7b37-45ec-a57d-ebb07978d6fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.167674 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-kube-api-access-kbrzr" (OuterVolumeSpecName: "kube-api-access-kbrzr") pod "1e820e3b-7b37-45ec-a57d-ebb07978d6fa" (UID: "1e820e3b-7b37-45ec-a57d-ebb07978d6fa"). InnerVolumeSpecName "kube-api-access-kbrzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.257820 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e820e3b-7b37-45ec-a57d-ebb07978d6fa" (UID: "1e820e3b-7b37-45ec-a57d-ebb07978d6fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.262210 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbrzr\" (UniqueName: \"kubernetes.io/projected/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-kube-api-access-kbrzr\") on node \"crc\" DevicePath \"\"" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.262245 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.262254 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e820e3b-7b37-45ec-a57d-ebb07978d6fa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.416442 4771 generic.go:334] "Generic (PLEG): container finished" podID="1e820e3b-7b37-45ec-a57d-ebb07978d6fa" containerID="0d63137206c8adab41354d0d591848fdf0211b5776ec2bb06faf62fb59ed875a" exitCode=0 Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.416520 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67xqh" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.416518 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67xqh" event={"ID":"1e820e3b-7b37-45ec-a57d-ebb07978d6fa","Type":"ContainerDied","Data":"0d63137206c8adab41354d0d591848fdf0211b5776ec2bb06faf62fb59ed875a"} Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.416657 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67xqh" event={"ID":"1e820e3b-7b37-45ec-a57d-ebb07978d6fa","Type":"ContainerDied","Data":"1aaca6614d6d6a89a5b9b35db40e5fda3947586306bd9392431861962bb6cd83"} Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.416677 4771 scope.go:117] "RemoveContainer" containerID="0d63137206c8adab41354d0d591848fdf0211b5776ec2bb06faf62fb59ed875a" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.446382 4771 scope.go:117] "RemoveContainer" containerID="126e26a50268b991e0d0c29ec5107b7517b4e5f51672ccb015e3be8525548203" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.469693 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-67xqh"] Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.486751 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-67xqh"] Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.508827 4771 scope.go:117] "RemoveContainer" containerID="dadad7faca2b248d2535ad7034b000a1126aa48c8784fea2ca273ddcee7890c8" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.565812 4771 scope.go:117] "RemoveContainer" containerID="0d63137206c8adab41354d0d591848fdf0211b5776ec2bb06faf62fb59ed875a" Oct 02 10:43:52 crc kubenswrapper[4771]: E1002 10:43:52.567573 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d63137206c8adab41354d0d591848fdf0211b5776ec2bb06faf62fb59ed875a\": container with ID starting with 0d63137206c8adab41354d0d591848fdf0211b5776ec2bb06faf62fb59ed875a not found: ID does not exist" containerID="0d63137206c8adab41354d0d591848fdf0211b5776ec2bb06faf62fb59ed875a" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.567656 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d63137206c8adab41354d0d591848fdf0211b5776ec2bb06faf62fb59ed875a"} err="failed to get container status \"0d63137206c8adab41354d0d591848fdf0211b5776ec2bb06faf62fb59ed875a\": rpc error: code = NotFound desc = could not find container \"0d63137206c8adab41354d0d591848fdf0211b5776ec2bb06faf62fb59ed875a\": container with ID starting with 0d63137206c8adab41354d0d591848fdf0211b5776ec2bb06faf62fb59ed875a not found: ID does not exist" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.567708 4771 scope.go:117] "RemoveContainer" containerID="126e26a50268b991e0d0c29ec5107b7517b4e5f51672ccb015e3be8525548203" Oct 02 10:43:52 crc kubenswrapper[4771]: E1002 10:43:52.568069 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"126e26a50268b991e0d0c29ec5107b7517b4e5f51672ccb015e3be8525548203\": container with ID starting with 126e26a50268b991e0d0c29ec5107b7517b4e5f51672ccb015e3be8525548203 not found: ID does not exist" containerID="126e26a50268b991e0d0c29ec5107b7517b4e5f51672ccb015e3be8525548203" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.568117 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"126e26a50268b991e0d0c29ec5107b7517b4e5f51672ccb015e3be8525548203"} err="failed to get container status \"126e26a50268b991e0d0c29ec5107b7517b4e5f51672ccb015e3be8525548203\": rpc error: code = NotFound desc = could not find container \"126e26a50268b991e0d0c29ec5107b7517b4e5f51672ccb015e3be8525548203\": container with ID starting with 126e26a50268b991e0d0c29ec5107b7517b4e5f51672ccb015e3be8525548203 not found: ID does not exist" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.568167 4771 scope.go:117] "RemoveContainer" containerID="dadad7faca2b248d2535ad7034b000a1126aa48c8784fea2ca273ddcee7890c8" Oct 02 10:43:52 crc kubenswrapper[4771]: E1002 10:43:52.568436 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dadad7faca2b248d2535ad7034b000a1126aa48c8784fea2ca273ddcee7890c8\": container with ID starting with dadad7faca2b248d2535ad7034b000a1126aa48c8784fea2ca273ddcee7890c8 not found: ID does not exist" containerID="dadad7faca2b248d2535ad7034b000a1126aa48c8784fea2ca273ddcee7890c8" Oct 02 10:43:52 crc kubenswrapper[4771]: I1002 10:43:52.568465 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dadad7faca2b248d2535ad7034b000a1126aa48c8784fea2ca273ddcee7890c8"} err="failed to get container status \"dadad7faca2b248d2535ad7034b000a1126aa48c8784fea2ca273ddcee7890c8\": rpc error: code = NotFound desc = could not find container \"dadad7faca2b248d2535ad7034b000a1126aa48c8784fea2ca273ddcee7890c8\": container with ID starting with dadad7faca2b248d2535ad7034b000a1126aa48c8784fea2ca273ddcee7890c8 not found: ID does not exist" Oct 02 10:43:53 crc kubenswrapper[4771]: I1002 10:43:53.704652 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e820e3b-7b37-45ec-a57d-ebb07978d6fa" path="/var/lib/kubelet/pods/1e820e3b-7b37-45ec-a57d-ebb07978d6fa/volumes" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.153595 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg"] Oct 02 10:45:00 crc kubenswrapper[4771]: E1002 10:45:00.155865 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e820e3b-7b37-45ec-a57d-ebb07978d6fa" containerName="extract-utilities" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.155965 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e820e3b-7b37-45ec-a57d-ebb07978d6fa" containerName="extract-utilities" Oct 02 10:45:00 crc kubenswrapper[4771]: E1002 10:45:00.156082 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e820e3b-7b37-45ec-a57d-ebb07978d6fa" containerName="registry-server" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.156172 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e820e3b-7b37-45ec-a57d-ebb07978d6fa" containerName="registry-server" Oct 02 10:45:00 crc kubenswrapper[4771]: E1002 10:45:00.156259 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e820e3b-7b37-45ec-a57d-ebb07978d6fa" containerName="extract-content" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.156423 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e820e3b-7b37-45ec-a57d-ebb07978d6fa" containerName="extract-content" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.156931 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e820e3b-7b37-45ec-a57d-ebb07978d6fa" containerName="registry-server" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.158440 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.162061 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.163295 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.209244 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg"] Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.293261 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-secret-volume\") pod \"collect-profiles-29323365-8nwzg\" (UID: \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.293670 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-config-volume\") pod \"collect-profiles-29323365-8nwzg\" (UID: \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.294231 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfpzk\" (UniqueName: \"kubernetes.io/projected/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-kube-api-access-sfpzk\") pod \"collect-profiles-29323365-8nwzg\" (UID: \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.397704 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfpzk\" (UniqueName: \"kubernetes.io/projected/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-kube-api-access-sfpzk\") pod \"collect-profiles-29323365-8nwzg\" (UID: \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.397864 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-secret-volume\") pod \"collect-profiles-29323365-8nwzg\" (UID: \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.397924 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-config-volume\") pod \"collect-profiles-29323365-8nwzg\" (UID: \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.399154 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-config-volume\") pod \"collect-profiles-29323365-8nwzg\" (UID: \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.405704 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-secret-volume\") pod \"collect-profiles-29323365-8nwzg\" (UID: \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.441292 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfpzk\" (UniqueName: \"kubernetes.io/projected/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-kube-api-access-sfpzk\") pod \"collect-profiles-29323365-8nwzg\" (UID: \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" Oct 02 10:45:00 crc kubenswrapper[4771]: I1002 10:45:00.521643 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" Oct 02 10:45:01 crc kubenswrapper[4771]: I1002 10:45:01.044594 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg"] Oct 02 10:45:01 crc kubenswrapper[4771]: I1002 10:45:01.189986 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" event={"ID":"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788","Type":"ContainerStarted","Data":"f96ac60217ddb59fcb43827a2dee011da64a56e478311112b88744f5f961f1c7"} Oct 02 10:45:02 crc kubenswrapper[4771]: I1002 10:45:02.204928 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" event={"ID":"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788","Type":"ContainerStarted","Data":"1a064682a845bcc5e3e48492a585aa00f2ce166c5bfdcd5315b2b3e4b2312760"} Oct 02 10:45:02 crc kubenswrapper[4771]: I1002 10:45:02.232479 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" podStartSLOduration=2.232456919 podStartE2EDuration="2.232456919s" podCreationTimestamp="2025-10-02 10:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:45:02.227393926 +0000 UTC m=+4089.875079003" watchObservedRunningTime="2025-10-02 10:45:02.232456919 +0000 UTC m=+4089.880141986" Oct 02 10:45:03 crc kubenswrapper[4771]: I1002 10:45:03.215939 4771 generic.go:334] "Generic (PLEG): container finished" podID="0ee7e9cd-4105-4ea8-bdfc-ab469fa68788" containerID="1a064682a845bcc5e3e48492a585aa00f2ce166c5bfdcd5315b2b3e4b2312760" exitCode=0 Oct 02 10:45:03 crc kubenswrapper[4771]: I1002 10:45:03.216288 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" event={"ID":"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788","Type":"ContainerDied","Data":"1a064682a845bcc5e3e48492a585aa00f2ce166c5bfdcd5315b2b3e4b2312760"} Oct 02 10:45:04 crc kubenswrapper[4771]: I1002 10:45:04.656008 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" Oct 02 10:45:04 crc kubenswrapper[4771]: I1002 10:45:04.707230 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-config-volume\") pod \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\" (UID: \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\") " Oct 02 10:45:04 crc kubenswrapper[4771]: I1002 10:45:04.707397 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-secret-volume\") pod \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\" (UID: \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\") " Oct 02 10:45:04 crc kubenswrapper[4771]: I1002 10:45:04.707533 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfpzk\" (UniqueName: \"kubernetes.io/projected/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-kube-api-access-sfpzk\") pod \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\" (UID: \"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788\") " Oct 02 10:45:04 crc kubenswrapper[4771]: I1002 10:45:04.708628 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-config-volume" (OuterVolumeSpecName: "config-volume") pod "0ee7e9cd-4105-4ea8-bdfc-ab469fa68788" (UID: "0ee7e9cd-4105-4ea8-bdfc-ab469fa68788"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:45:04 crc kubenswrapper[4771]: I1002 10:45:04.714097 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0ee7e9cd-4105-4ea8-bdfc-ab469fa68788" (UID: "0ee7e9cd-4105-4ea8-bdfc-ab469fa68788"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:45:04 crc kubenswrapper[4771]: I1002 10:45:04.717157 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-kube-api-access-sfpzk" (OuterVolumeSpecName: "kube-api-access-sfpzk") pod "0ee7e9cd-4105-4ea8-bdfc-ab469fa68788" (UID: "0ee7e9cd-4105-4ea8-bdfc-ab469fa68788"). InnerVolumeSpecName "kube-api-access-sfpzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:45:04 crc kubenswrapper[4771]: I1002 10:45:04.811667 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfpzk\" (UniqueName: \"kubernetes.io/projected/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-kube-api-access-sfpzk\") on node \"crc\" DevicePath \"\"" Oct 02 10:45:04 crc kubenswrapper[4771]: I1002 10:45:04.811727 4771 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:45:04 crc kubenswrapper[4771]: I1002 10:45:04.811739 4771 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:45:05 crc kubenswrapper[4771]: I1002 10:45:05.243813 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" event={"ID":"0ee7e9cd-4105-4ea8-bdfc-ab469fa68788","Type":"ContainerDied","Data":"f96ac60217ddb59fcb43827a2dee011da64a56e478311112b88744f5f961f1c7"} Oct 02 10:45:05 crc kubenswrapper[4771]: I1002 10:45:05.244107 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f96ac60217ddb59fcb43827a2dee011da64a56e478311112b88744f5f961f1c7" Oct 02 10:45:05 crc kubenswrapper[4771]: I1002 10:45:05.244171 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg" Oct 02 10:45:05 crc kubenswrapper[4771]: I1002 10:45:05.313018 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7"] Oct 02 10:45:05 crc kubenswrapper[4771]: I1002 10:45:05.324765 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-gqsw7"] Oct 02 10:45:05 crc kubenswrapper[4771]: I1002 10:45:05.699341 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="325b9e47-f103-41d4-ad1f-f9fb2eecde45" path="/var/lib/kubelet/pods/325b9e47-f103-41d4-ad1f-f9fb2eecde45/volumes" Oct 02 10:45:12 crc kubenswrapper[4771]: I1002 10:45:12.145939 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:45:12 crc kubenswrapper[4771]: I1002 10:45:12.146517 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:45:42 crc kubenswrapper[4771]: I1002 10:45:42.146184 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:45:42 crc kubenswrapper[4771]: I1002 10:45:42.146823 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:46:05 crc kubenswrapper[4771]: I1002 10:46:05.418549 4771 scope.go:117] "RemoveContainer" containerID="e138a2ca1b3d6a52d4b4f986aa194b1a52fe0f83d5004be678043c565d1242dd" Oct 02 10:46:12 crc kubenswrapper[4771]: I1002 10:46:12.146099 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:46:12 crc kubenswrapper[4771]: I1002 10:46:12.146682 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:46:12 crc kubenswrapper[4771]: I1002 10:46:12.146729 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 10:46:12 crc kubenswrapper[4771]: I1002 10:46:12.147935 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"194c5b729f6dfb0975af45a7ef28aedac3730958132074ff9a427ab736eddfcb"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:46:12 crc kubenswrapper[4771]: I1002 10:46:12.147988 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://194c5b729f6dfb0975af45a7ef28aedac3730958132074ff9a427ab736eddfcb" gracePeriod=600 Oct 02 10:46:12 crc kubenswrapper[4771]: I1002 10:46:12.972201 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="194c5b729f6dfb0975af45a7ef28aedac3730958132074ff9a427ab736eddfcb" exitCode=0 Oct 02 10:46:12 crc kubenswrapper[4771]: I1002 10:46:12.972334 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"194c5b729f6dfb0975af45a7ef28aedac3730958132074ff9a427ab736eddfcb"} Oct 02 10:46:12 crc kubenswrapper[4771]: I1002 10:46:12.972859 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764"} Oct 02 10:46:12 crc kubenswrapper[4771]: I1002 10:46:12.972888 4771 scope.go:117] "RemoveContainer" containerID="aba5ea85f65b6e4c45a8f41816b03e749d18689046e07f545332db418d7f4ecf" Oct 02 10:48:11 crc kubenswrapper[4771]: I1002 10:48:11.861487 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ng9hr"] Oct 02 10:48:11 crc kubenswrapper[4771]: E1002 10:48:11.862614 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ee7e9cd-4105-4ea8-bdfc-ab469fa68788" containerName="collect-profiles" Oct 02 10:48:11 crc kubenswrapper[4771]: I1002 10:48:11.862630 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ee7e9cd-4105-4ea8-bdfc-ab469fa68788" containerName="collect-profiles" Oct 02 10:48:11 crc kubenswrapper[4771]: I1002 10:48:11.863012 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ee7e9cd-4105-4ea8-bdfc-ab469fa68788" containerName="collect-profiles" Oct 02 10:48:11 crc kubenswrapper[4771]: I1002 10:48:11.865356 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:11 crc kubenswrapper[4771]: I1002 10:48:11.876905 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ng9hr"] Oct 02 10:48:11 crc kubenswrapper[4771]: I1002 10:48:11.965479 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-utilities\") pod \"redhat-marketplace-ng9hr\" (UID: \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\") " pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:11 crc kubenswrapper[4771]: I1002 10:48:11.965741 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-catalog-content\") pod \"redhat-marketplace-ng9hr\" (UID: \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\") " pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:11 crc kubenswrapper[4771]: I1002 10:48:11.965858 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjw77\" (UniqueName: \"kubernetes.io/projected/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-kube-api-access-wjw77\") pod \"redhat-marketplace-ng9hr\" (UID: \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\") " pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:12 crc kubenswrapper[4771]: I1002 10:48:12.068681 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-catalog-content\") pod \"redhat-marketplace-ng9hr\" (UID: \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\") " pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:12 crc kubenswrapper[4771]: I1002 10:48:12.068786 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjw77\" (UniqueName: \"kubernetes.io/projected/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-kube-api-access-wjw77\") pod \"redhat-marketplace-ng9hr\" (UID: \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\") " pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:12 crc kubenswrapper[4771]: I1002 10:48:12.068963 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-utilities\") pod \"redhat-marketplace-ng9hr\" (UID: \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\") " pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:12 crc kubenswrapper[4771]: I1002 10:48:12.069248 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-catalog-content\") pod \"redhat-marketplace-ng9hr\" (UID: \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\") " pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:12 crc kubenswrapper[4771]: I1002 10:48:12.069550 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-utilities\") pod \"redhat-marketplace-ng9hr\" (UID: \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\") " pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:12 crc kubenswrapper[4771]: I1002 10:48:12.146538 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:48:12 crc kubenswrapper[4771]: I1002 10:48:12.146628 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:48:12 crc kubenswrapper[4771]: I1002 10:48:12.474662 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjw77\" (UniqueName: \"kubernetes.io/projected/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-kube-api-access-wjw77\") pod \"redhat-marketplace-ng9hr\" (UID: \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\") " pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:12 crc kubenswrapper[4771]: I1002 10:48:12.502631 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:12 crc kubenswrapper[4771]: I1002 10:48:12.997222 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ng9hr"] Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.284505 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ng9hr" event={"ID":"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5","Type":"ContainerStarted","Data":"9e2156b3630a7afad0ce7357c321054e4ca7519a98ca1fbd5fdb9de8f5a610d4"} Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.284810 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ng9hr" event={"ID":"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5","Type":"ContainerStarted","Data":"9a4be9196b7f6ed126b7ae88565f6ee6bd264b8322a1c1b96a1b6b609acd8487"} Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.672671 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vl4bk"] Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.676019 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.752471 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vl4bk"] Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.813600 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-utilities\") pod \"certified-operators-vl4bk\" (UID: \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\") " pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.813774 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-catalog-content\") pod \"certified-operators-vl4bk\" (UID: \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\") " pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.813808 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pngnn\" (UniqueName: \"kubernetes.io/projected/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-kube-api-access-pngnn\") pod \"certified-operators-vl4bk\" (UID: \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\") " pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.916967 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-utilities\") pod \"certified-operators-vl4bk\" (UID: \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\") " pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.917052 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-catalog-content\") pod \"certified-operators-vl4bk\" (UID: \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\") " pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.917073 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pngnn\" (UniqueName: \"kubernetes.io/projected/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-kube-api-access-pngnn\") pod \"certified-operators-vl4bk\" (UID: \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\") " pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.919194 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-utilities\") pod \"certified-operators-vl4bk\" (UID: \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\") " pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.919432 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-catalog-content\") pod \"certified-operators-vl4bk\" (UID: \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\") " pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:13 crc kubenswrapper[4771]: I1002 10:48:13.959861 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pngnn\" (UniqueName: \"kubernetes.io/projected/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-kube-api-access-pngnn\") pod \"certified-operators-vl4bk\" (UID: \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\") " pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:14 crc kubenswrapper[4771]: I1002 10:48:14.046324 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:14 crc kubenswrapper[4771]: I1002 10:48:14.298101 4771 generic.go:334] "Generic (PLEG): container finished" podID="6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" containerID="9e2156b3630a7afad0ce7357c321054e4ca7519a98ca1fbd5fdb9de8f5a610d4" exitCode=0 Oct 02 10:48:14 crc kubenswrapper[4771]: I1002 10:48:14.298187 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ng9hr" event={"ID":"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5","Type":"ContainerDied","Data":"9e2156b3630a7afad0ce7357c321054e4ca7519a98ca1fbd5fdb9de8f5a610d4"} Oct 02 10:48:14 crc kubenswrapper[4771]: I1002 10:48:14.610706 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vl4bk"] Oct 02 10:48:14 crc kubenswrapper[4771]: W1002 10:48:14.616020 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod869d9a08_35b0_4cf9_97a6_b5939a32ffb4.slice/crio-4aed234fb3183c63001884a824c24c877762e9307aebf0204da6015e56415696 WatchSource:0}: Error finding container 4aed234fb3183c63001884a824c24c877762e9307aebf0204da6015e56415696: Status 404 returned error can't find the container with id 4aed234fb3183c63001884a824c24c877762e9307aebf0204da6015e56415696 Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.057656 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vkskd"] Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.060792 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.079644 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vkskd"] Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.155213 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bst8z\" (UniqueName: \"kubernetes.io/projected/f1ece6dc-bff0-436d-a81b-8963a04162e9-kube-api-access-bst8z\") pod \"community-operators-vkskd\" (UID: \"f1ece6dc-bff0-436d-a81b-8963a04162e9\") " pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.155398 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1ece6dc-bff0-436d-a81b-8963a04162e9-utilities\") pod \"community-operators-vkskd\" (UID: \"f1ece6dc-bff0-436d-a81b-8963a04162e9\") " pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.155464 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1ece6dc-bff0-436d-a81b-8963a04162e9-catalog-content\") pod \"community-operators-vkskd\" (UID: \"f1ece6dc-bff0-436d-a81b-8963a04162e9\") " pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.258512 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bst8z\" (UniqueName: \"kubernetes.io/projected/f1ece6dc-bff0-436d-a81b-8963a04162e9-kube-api-access-bst8z\") pod \"community-operators-vkskd\" (UID: \"f1ece6dc-bff0-436d-a81b-8963a04162e9\") " pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.258607 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1ece6dc-bff0-436d-a81b-8963a04162e9-utilities\") pod \"community-operators-vkskd\" (UID: \"f1ece6dc-bff0-436d-a81b-8963a04162e9\") " pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.258655 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1ece6dc-bff0-436d-a81b-8963a04162e9-catalog-content\") pod \"community-operators-vkskd\" (UID: \"f1ece6dc-bff0-436d-a81b-8963a04162e9\") " pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.259345 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1ece6dc-bff0-436d-a81b-8963a04162e9-utilities\") pod \"community-operators-vkskd\" (UID: \"f1ece6dc-bff0-436d-a81b-8963a04162e9\") " pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.259497 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1ece6dc-bff0-436d-a81b-8963a04162e9-catalog-content\") pod \"community-operators-vkskd\" (UID: \"f1ece6dc-bff0-436d-a81b-8963a04162e9\") " pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.282371 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bst8z\" (UniqueName: \"kubernetes.io/projected/f1ece6dc-bff0-436d-a81b-8963a04162e9-kube-api-access-bst8z\") pod \"community-operators-vkskd\" (UID: \"f1ece6dc-bff0-436d-a81b-8963a04162e9\") " pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.312254 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl4bk" event={"ID":"869d9a08-35b0-4cf9-97a6-b5939a32ffb4","Type":"ContainerDied","Data":"4cfd5d3631c50c0285169fe1bac4c23a5ef6aacacbd59b3190cac6b1d110330e"} Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.312442 4771 generic.go:334] "Generic (PLEG): container finished" podID="869d9a08-35b0-4cf9-97a6-b5939a32ffb4" containerID="4cfd5d3631c50c0285169fe1bac4c23a5ef6aacacbd59b3190cac6b1d110330e" exitCode=0 Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.312566 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl4bk" event={"ID":"869d9a08-35b0-4cf9-97a6-b5939a32ffb4","Type":"ContainerStarted","Data":"4aed234fb3183c63001884a824c24c877762e9307aebf0204da6015e56415696"} Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.322946 4771 generic.go:334] "Generic (PLEG): container finished" podID="6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" containerID="70d52cd714c7210bdc17d02c2a2ffad4113b731c707c41854998cbc8719e80d3" exitCode=0 Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.323501 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ng9hr" event={"ID":"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5","Type":"ContainerDied","Data":"70d52cd714c7210bdc17d02c2a2ffad4113b731c707c41854998cbc8719e80d3"} Oct 02 10:48:15 crc kubenswrapper[4771]: I1002 10:48:15.392081 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:16 crc kubenswrapper[4771]: I1002 10:48:16.006039 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vkskd"] Oct 02 10:48:16 crc kubenswrapper[4771]: I1002 10:48:16.340481 4771 generic.go:334] "Generic (PLEG): container finished" podID="f1ece6dc-bff0-436d-a81b-8963a04162e9" containerID="e74df73d8ce599ec6dc23cc26366c4f678a916f72a21dd0b3ee37e0440649efe" exitCode=0 Oct 02 10:48:16 crc kubenswrapper[4771]: I1002 10:48:16.340556 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkskd" event={"ID":"f1ece6dc-bff0-436d-a81b-8963a04162e9","Type":"ContainerDied","Data":"e74df73d8ce599ec6dc23cc26366c4f678a916f72a21dd0b3ee37e0440649efe"} Oct 02 10:48:16 crc kubenswrapper[4771]: I1002 10:48:16.340805 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkskd" event={"ID":"f1ece6dc-bff0-436d-a81b-8963a04162e9","Type":"ContainerStarted","Data":"c04e036dfc04be098202835e60dac8f5f29223fed9cfaab5714e54ced2359d85"} Oct 02 10:48:17 crc kubenswrapper[4771]: I1002 10:48:17.354636 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl4bk" event={"ID":"869d9a08-35b0-4cf9-97a6-b5939a32ffb4","Type":"ContainerStarted","Data":"19067b26d87b406b3912303402436a1248285b16beca484950a55f5a540ff60d"} Oct 02 10:48:17 crc kubenswrapper[4771]: I1002 10:48:17.360359 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ng9hr" event={"ID":"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5","Type":"ContainerStarted","Data":"17c2290ff602761eedda9306ab8e1702792291367e93106620e231b8422158cc"} Oct 02 10:48:17 crc kubenswrapper[4771]: I1002 10:48:17.433513 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ng9hr" podStartSLOduration=3.395981701 podStartE2EDuration="6.433487701s" podCreationTimestamp="2025-10-02 10:48:11 +0000 UTC" firstStartedPulling="2025-10-02 10:48:13.28675829 +0000 UTC m=+4280.934443347" lastFinishedPulling="2025-10-02 10:48:16.32426428 +0000 UTC m=+4283.971949347" observedRunningTime="2025-10-02 10:48:17.417673799 +0000 UTC m=+4285.065358876" watchObservedRunningTime="2025-10-02 10:48:17.433487701 +0000 UTC m=+4285.081172768" Oct 02 10:48:18 crc kubenswrapper[4771]: I1002 10:48:18.379867 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkskd" event={"ID":"f1ece6dc-bff0-436d-a81b-8963a04162e9","Type":"ContainerStarted","Data":"0d070db56d170aa762f1b190c3e84f7efc974e30e582ec7808ec7d8bfb6912bc"} Oct 02 10:48:20 crc kubenswrapper[4771]: I1002 10:48:20.402184 4771 generic.go:334] "Generic (PLEG): container finished" podID="869d9a08-35b0-4cf9-97a6-b5939a32ffb4" containerID="19067b26d87b406b3912303402436a1248285b16beca484950a55f5a540ff60d" exitCode=0 Oct 02 10:48:20 crc kubenswrapper[4771]: I1002 10:48:20.402253 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl4bk" event={"ID":"869d9a08-35b0-4cf9-97a6-b5939a32ffb4","Type":"ContainerDied","Data":"19067b26d87b406b3912303402436a1248285b16beca484950a55f5a540ff60d"} Oct 02 10:48:20 crc kubenswrapper[4771]: I1002 10:48:20.406647 4771 generic.go:334] "Generic (PLEG): container finished" podID="f1ece6dc-bff0-436d-a81b-8963a04162e9" containerID="0d070db56d170aa762f1b190c3e84f7efc974e30e582ec7808ec7d8bfb6912bc" exitCode=0 Oct 02 10:48:20 crc kubenswrapper[4771]: I1002 10:48:20.406757 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkskd" event={"ID":"f1ece6dc-bff0-436d-a81b-8963a04162e9","Type":"ContainerDied","Data":"0d070db56d170aa762f1b190c3e84f7efc974e30e582ec7808ec7d8bfb6912bc"} Oct 02 10:48:21 crc kubenswrapper[4771]: I1002 10:48:21.419838 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkskd" event={"ID":"f1ece6dc-bff0-436d-a81b-8963a04162e9","Type":"ContainerStarted","Data":"38c693208bc6538300642f24abd38144d8f2ca6a08dbe9a7367662f9388ce6a6"} Oct 02 10:48:21 crc kubenswrapper[4771]: I1002 10:48:21.426648 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl4bk" event={"ID":"869d9a08-35b0-4cf9-97a6-b5939a32ffb4","Type":"ContainerStarted","Data":"f684cd2c5a2a9cd252c855af0badf63399158f51f63e2680f3ece44b11d82dba"} Oct 02 10:48:21 crc kubenswrapper[4771]: I1002 10:48:21.452042 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vkskd" podStartSLOduration=1.86853369 podStartE2EDuration="6.452024492s" podCreationTimestamp="2025-10-02 10:48:15 +0000 UTC" firstStartedPulling="2025-10-02 10:48:16.353816329 +0000 UTC m=+4284.001501396" lastFinishedPulling="2025-10-02 10:48:20.937307131 +0000 UTC m=+4288.584992198" observedRunningTime="2025-10-02 10:48:21.444574893 +0000 UTC m=+4289.092260160" watchObservedRunningTime="2025-10-02 10:48:21.452024492 +0000 UTC m=+4289.099709549" Oct 02 10:48:21 crc kubenswrapper[4771]: I1002 10:48:21.467168 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vl4bk" podStartSLOduration=2.930400674 podStartE2EDuration="8.467147636s" podCreationTimestamp="2025-10-02 10:48:13 +0000 UTC" firstStartedPulling="2025-10-02 10:48:15.314535704 +0000 UTC m=+4282.962220771" lastFinishedPulling="2025-10-02 10:48:20.851282666 +0000 UTC m=+4288.498967733" observedRunningTime="2025-10-02 10:48:21.462586854 +0000 UTC m=+4289.110271921" watchObservedRunningTime="2025-10-02 10:48:21.467147636 +0000 UTC m=+4289.114832713" Oct 02 10:48:22 crc kubenswrapper[4771]: I1002 10:48:22.503765 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:22 crc kubenswrapper[4771]: I1002 10:48:22.503843 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:22 crc kubenswrapper[4771]: I1002 10:48:22.557031 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:23 crc kubenswrapper[4771]: I1002 10:48:23.497056 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:24 crc kubenswrapper[4771]: I1002 10:48:24.046550 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:24 crc kubenswrapper[4771]: I1002 10:48:24.047504 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:25 crc kubenswrapper[4771]: I1002 10:48:25.098191 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-vl4bk" podUID="869d9a08-35b0-4cf9-97a6-b5939a32ffb4" containerName="registry-server" probeResult="failure" output=< Oct 02 10:48:25 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:48:25 crc kubenswrapper[4771]: > Oct 02 10:48:25 crc kubenswrapper[4771]: I1002 10:48:25.393020 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:25 crc kubenswrapper[4771]: I1002 10:48:25.393867 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:25 crc kubenswrapper[4771]: I1002 10:48:25.449563 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ng9hr"] Oct 02 10:48:25 crc kubenswrapper[4771]: I1002 10:48:25.471238 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ng9hr" podUID="6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" containerName="registry-server" containerID="cri-o://17c2290ff602761eedda9306ab8e1702792291367e93106620e231b8422158cc" gracePeriod=2 Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.005155 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.156001 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjw77\" (UniqueName: \"kubernetes.io/projected/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-kube-api-access-wjw77\") pod \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\" (UID: \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\") " Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.156629 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-utilities\") pod \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\" (UID: \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\") " Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.156879 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-catalog-content\") pod \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\" (UID: \"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5\") " Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.157168 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-utilities" (OuterVolumeSpecName: "utilities") pod "6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" (UID: "6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.157387 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.163526 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-kube-api-access-wjw77" (OuterVolumeSpecName: "kube-api-access-wjw77") pod "6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" (UID: "6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5"). InnerVolumeSpecName "kube-api-access-wjw77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.170655 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" (UID: "6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.259769 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.259807 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjw77\" (UniqueName: \"kubernetes.io/projected/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5-kube-api-access-wjw77\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.449086 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-vkskd" podUID="f1ece6dc-bff0-436d-a81b-8963a04162e9" containerName="registry-server" probeResult="failure" output=< Oct 02 10:48:26 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:48:26 crc kubenswrapper[4771]: > Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.483915 4771 generic.go:334] "Generic (PLEG): container finished" podID="6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" containerID="17c2290ff602761eedda9306ab8e1702792291367e93106620e231b8422158cc" exitCode=0 Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.483966 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ng9hr" event={"ID":"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5","Type":"ContainerDied","Data":"17c2290ff602761eedda9306ab8e1702792291367e93106620e231b8422158cc"} Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.483999 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ng9hr" event={"ID":"6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5","Type":"ContainerDied","Data":"9a4be9196b7f6ed126b7ae88565f6ee6bd264b8322a1c1b96a1b6b609acd8487"} Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.484022 4771 scope.go:117] "RemoveContainer" containerID="17c2290ff602761eedda9306ab8e1702792291367e93106620e231b8422158cc" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.484077 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ng9hr" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.527882 4771 scope.go:117] "RemoveContainer" containerID="70d52cd714c7210bdc17d02c2a2ffad4113b731c707c41854998cbc8719e80d3" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.530240 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ng9hr"] Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.543216 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ng9hr"] Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.554280 4771 scope.go:117] "RemoveContainer" containerID="9e2156b3630a7afad0ce7357c321054e4ca7519a98ca1fbd5fdb9de8f5a610d4" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.616498 4771 scope.go:117] "RemoveContainer" containerID="17c2290ff602761eedda9306ab8e1702792291367e93106620e231b8422158cc" Oct 02 10:48:26 crc kubenswrapper[4771]: E1002 10:48:26.617305 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17c2290ff602761eedda9306ab8e1702792291367e93106620e231b8422158cc\": container with ID starting with 17c2290ff602761eedda9306ab8e1702792291367e93106620e231b8422158cc not found: ID does not exist" containerID="17c2290ff602761eedda9306ab8e1702792291367e93106620e231b8422158cc" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.617345 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17c2290ff602761eedda9306ab8e1702792291367e93106620e231b8422158cc"} err="failed to get container status \"17c2290ff602761eedda9306ab8e1702792291367e93106620e231b8422158cc\": rpc error: code = NotFound desc = could not find container \"17c2290ff602761eedda9306ab8e1702792291367e93106620e231b8422158cc\": container with ID starting with 17c2290ff602761eedda9306ab8e1702792291367e93106620e231b8422158cc not found: ID does not exist" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.617375 4771 scope.go:117] "RemoveContainer" containerID="70d52cd714c7210bdc17d02c2a2ffad4113b731c707c41854998cbc8719e80d3" Oct 02 10:48:26 crc kubenswrapper[4771]: E1002 10:48:26.617669 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70d52cd714c7210bdc17d02c2a2ffad4113b731c707c41854998cbc8719e80d3\": container with ID starting with 70d52cd714c7210bdc17d02c2a2ffad4113b731c707c41854998cbc8719e80d3 not found: ID does not exist" containerID="70d52cd714c7210bdc17d02c2a2ffad4113b731c707c41854998cbc8719e80d3" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.617912 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70d52cd714c7210bdc17d02c2a2ffad4113b731c707c41854998cbc8719e80d3"} err="failed to get container status \"70d52cd714c7210bdc17d02c2a2ffad4113b731c707c41854998cbc8719e80d3\": rpc error: code = NotFound desc = could not find container \"70d52cd714c7210bdc17d02c2a2ffad4113b731c707c41854998cbc8719e80d3\": container with ID starting with 70d52cd714c7210bdc17d02c2a2ffad4113b731c707c41854998cbc8719e80d3 not found: ID does not exist" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.617948 4771 scope.go:117] "RemoveContainer" containerID="9e2156b3630a7afad0ce7357c321054e4ca7519a98ca1fbd5fdb9de8f5a610d4" Oct 02 10:48:26 crc kubenswrapper[4771]: E1002 10:48:26.618377 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e2156b3630a7afad0ce7357c321054e4ca7519a98ca1fbd5fdb9de8f5a610d4\": container with ID starting with 9e2156b3630a7afad0ce7357c321054e4ca7519a98ca1fbd5fdb9de8f5a610d4 not found: ID does not exist" containerID="9e2156b3630a7afad0ce7357c321054e4ca7519a98ca1fbd5fdb9de8f5a610d4" Oct 02 10:48:26 crc kubenswrapper[4771]: I1002 10:48:26.618425 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e2156b3630a7afad0ce7357c321054e4ca7519a98ca1fbd5fdb9de8f5a610d4"} err="failed to get container status \"9e2156b3630a7afad0ce7357c321054e4ca7519a98ca1fbd5fdb9de8f5a610d4\": rpc error: code = NotFound desc = could not find container \"9e2156b3630a7afad0ce7357c321054e4ca7519a98ca1fbd5fdb9de8f5a610d4\": container with ID starting with 9e2156b3630a7afad0ce7357c321054e4ca7519a98ca1fbd5fdb9de8f5a610d4 not found: ID does not exist" Oct 02 10:48:27 crc kubenswrapper[4771]: I1002 10:48:27.694852 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" path="/var/lib/kubelet/pods/6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5/volumes" Oct 02 10:48:34 crc kubenswrapper[4771]: I1002 10:48:34.100848 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:34 crc kubenswrapper[4771]: I1002 10:48:34.154923 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:34 crc kubenswrapper[4771]: I1002 10:48:34.343239 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vl4bk"] Oct 02 10:48:35 crc kubenswrapper[4771]: I1002 10:48:35.577888 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vl4bk" podUID="869d9a08-35b0-4cf9-97a6-b5939a32ffb4" containerName="registry-server" containerID="cri-o://f684cd2c5a2a9cd252c855af0badf63399158f51f63e2680f3ece44b11d82dba" gracePeriod=2 Oct 02 10:48:36 crc kubenswrapper[4771]: I1002 10:48:36.438883 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-vkskd" podUID="f1ece6dc-bff0-436d-a81b-8963a04162e9" containerName="registry-server" probeResult="failure" output=< Oct 02 10:48:36 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:48:36 crc kubenswrapper[4771]: > Oct 02 10:48:36 crc kubenswrapper[4771]: I1002 10:48:36.592389 4771 generic.go:334] "Generic (PLEG): container finished" podID="869d9a08-35b0-4cf9-97a6-b5939a32ffb4" containerID="f684cd2c5a2a9cd252c855af0badf63399158f51f63e2680f3ece44b11d82dba" exitCode=0 Oct 02 10:48:36 crc kubenswrapper[4771]: I1002 10:48:36.592868 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl4bk" event={"ID":"869d9a08-35b0-4cf9-97a6-b5939a32ffb4","Type":"ContainerDied","Data":"f684cd2c5a2a9cd252c855af0badf63399158f51f63e2680f3ece44b11d82dba"} Oct 02 10:48:36 crc kubenswrapper[4771]: I1002 10:48:36.759280 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:36 crc kubenswrapper[4771]: I1002 10:48:36.856962 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pngnn\" (UniqueName: \"kubernetes.io/projected/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-kube-api-access-pngnn\") pod \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\" (UID: \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\") " Oct 02 10:48:36 crc kubenswrapper[4771]: I1002 10:48:36.857296 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-utilities\") pod \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\" (UID: \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\") " Oct 02 10:48:36 crc kubenswrapper[4771]: I1002 10:48:36.857327 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-catalog-content\") pod \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\" (UID: \"869d9a08-35b0-4cf9-97a6-b5939a32ffb4\") " Oct 02 10:48:36 crc kubenswrapper[4771]: I1002 10:48:36.860629 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-utilities" (OuterVolumeSpecName: "utilities") pod "869d9a08-35b0-4cf9-97a6-b5939a32ffb4" (UID: "869d9a08-35b0-4cf9-97a6-b5939a32ffb4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:48:36 crc kubenswrapper[4771]: I1002 10:48:36.868470 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-kube-api-access-pngnn" (OuterVolumeSpecName: "kube-api-access-pngnn") pod "869d9a08-35b0-4cf9-97a6-b5939a32ffb4" (UID: "869d9a08-35b0-4cf9-97a6-b5939a32ffb4"). InnerVolumeSpecName "kube-api-access-pngnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:48:36 crc kubenswrapper[4771]: I1002 10:48:36.909219 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "869d9a08-35b0-4cf9-97a6-b5939a32ffb4" (UID: "869d9a08-35b0-4cf9-97a6-b5939a32ffb4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:48:36 crc kubenswrapper[4771]: I1002 10:48:36.960408 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:36 crc kubenswrapper[4771]: I1002 10:48:36.960441 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:36 crc kubenswrapper[4771]: I1002 10:48:36.960454 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pngnn\" (UniqueName: \"kubernetes.io/projected/869d9a08-35b0-4cf9-97a6-b5939a32ffb4-kube-api-access-pngnn\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:37 crc kubenswrapper[4771]: I1002 10:48:37.614747 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vl4bk" event={"ID":"869d9a08-35b0-4cf9-97a6-b5939a32ffb4","Type":"ContainerDied","Data":"4aed234fb3183c63001884a824c24c877762e9307aebf0204da6015e56415696"} Oct 02 10:48:37 crc kubenswrapper[4771]: I1002 10:48:37.614800 4771 scope.go:117] "RemoveContainer" containerID="f684cd2c5a2a9cd252c855af0badf63399158f51f63e2680f3ece44b11d82dba" Oct 02 10:48:37 crc kubenswrapper[4771]: I1002 10:48:37.614943 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vl4bk" Oct 02 10:48:37 crc kubenswrapper[4771]: I1002 10:48:37.644188 4771 scope.go:117] "RemoveContainer" containerID="19067b26d87b406b3912303402436a1248285b16beca484950a55f5a540ff60d" Oct 02 10:48:37 crc kubenswrapper[4771]: I1002 10:48:37.666365 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vl4bk"] Oct 02 10:48:37 crc kubenswrapper[4771]: I1002 10:48:37.677084 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vl4bk"] Oct 02 10:48:37 crc kubenswrapper[4771]: I1002 10:48:37.695029 4771 scope.go:117] "RemoveContainer" containerID="4cfd5d3631c50c0285169fe1bac4c23a5ef6aacacbd59b3190cac6b1d110330e" Oct 02 10:48:37 crc kubenswrapper[4771]: I1002 10:48:37.704691 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="869d9a08-35b0-4cf9-97a6-b5939a32ffb4" path="/var/lib/kubelet/pods/869d9a08-35b0-4cf9-97a6-b5939a32ffb4/volumes" Oct 02 10:48:42 crc kubenswrapper[4771]: I1002 10:48:42.145986 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:48:42 crc kubenswrapper[4771]: I1002 10:48:42.146533 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:48:45 crc kubenswrapper[4771]: I1002 10:48:45.448371 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:45 crc kubenswrapper[4771]: I1002 10:48:45.506898 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:46 crc kubenswrapper[4771]: I1002 10:48:46.261648 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vkskd"] Oct 02 10:48:46 crc kubenswrapper[4771]: I1002 10:48:46.719867 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vkskd" podUID="f1ece6dc-bff0-436d-a81b-8963a04162e9" containerName="registry-server" containerID="cri-o://38c693208bc6538300642f24abd38144d8f2ca6a08dbe9a7367662f9388ce6a6" gracePeriod=2 Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.285139 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.427740 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bst8z\" (UniqueName: \"kubernetes.io/projected/f1ece6dc-bff0-436d-a81b-8963a04162e9-kube-api-access-bst8z\") pod \"f1ece6dc-bff0-436d-a81b-8963a04162e9\" (UID: \"f1ece6dc-bff0-436d-a81b-8963a04162e9\") " Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.427849 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1ece6dc-bff0-436d-a81b-8963a04162e9-catalog-content\") pod \"f1ece6dc-bff0-436d-a81b-8963a04162e9\" (UID: \"f1ece6dc-bff0-436d-a81b-8963a04162e9\") " Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.428030 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1ece6dc-bff0-436d-a81b-8963a04162e9-utilities\") pod \"f1ece6dc-bff0-436d-a81b-8963a04162e9\" (UID: \"f1ece6dc-bff0-436d-a81b-8963a04162e9\") " Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.428829 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1ece6dc-bff0-436d-a81b-8963a04162e9-utilities" (OuterVolumeSpecName: "utilities") pod "f1ece6dc-bff0-436d-a81b-8963a04162e9" (UID: "f1ece6dc-bff0-436d-a81b-8963a04162e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.429455 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1ece6dc-bff0-436d-a81b-8963a04162e9-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.441187 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1ece6dc-bff0-436d-a81b-8963a04162e9-kube-api-access-bst8z" (OuterVolumeSpecName: "kube-api-access-bst8z") pod "f1ece6dc-bff0-436d-a81b-8963a04162e9" (UID: "f1ece6dc-bff0-436d-a81b-8963a04162e9"). InnerVolumeSpecName "kube-api-access-bst8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.488335 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1ece6dc-bff0-436d-a81b-8963a04162e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1ece6dc-bff0-436d-a81b-8963a04162e9" (UID: "f1ece6dc-bff0-436d-a81b-8963a04162e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.531919 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bst8z\" (UniqueName: \"kubernetes.io/projected/f1ece6dc-bff0-436d-a81b-8963a04162e9-kube-api-access-bst8z\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.531963 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1ece6dc-bff0-436d-a81b-8963a04162e9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.733749 4771 generic.go:334] "Generic (PLEG): container finished" podID="f1ece6dc-bff0-436d-a81b-8963a04162e9" containerID="38c693208bc6538300642f24abd38144d8f2ca6a08dbe9a7367662f9388ce6a6" exitCode=0 Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.733807 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkskd" event={"ID":"f1ece6dc-bff0-436d-a81b-8963a04162e9","Type":"ContainerDied","Data":"38c693208bc6538300642f24abd38144d8f2ca6a08dbe9a7367662f9388ce6a6"} Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.733857 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkskd" event={"ID":"f1ece6dc-bff0-436d-a81b-8963a04162e9","Type":"ContainerDied","Data":"c04e036dfc04be098202835e60dac8f5f29223fed9cfaab5714e54ced2359d85"} Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.733880 4771 scope.go:117] "RemoveContainer" containerID="38c693208bc6538300642f24abd38144d8f2ca6a08dbe9a7367662f9388ce6a6" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.734071 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vkskd" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.766838 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vkskd"] Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.769831 4771 scope.go:117] "RemoveContainer" containerID="0d070db56d170aa762f1b190c3e84f7efc974e30e582ec7808ec7d8bfb6912bc" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.780977 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vkskd"] Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.793567 4771 scope.go:117] "RemoveContainer" containerID="e74df73d8ce599ec6dc23cc26366c4f678a916f72a21dd0b3ee37e0440649efe" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.852515 4771 scope.go:117] "RemoveContainer" containerID="38c693208bc6538300642f24abd38144d8f2ca6a08dbe9a7367662f9388ce6a6" Oct 02 10:48:47 crc kubenswrapper[4771]: E1002 10:48:47.853405 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38c693208bc6538300642f24abd38144d8f2ca6a08dbe9a7367662f9388ce6a6\": container with ID starting with 38c693208bc6538300642f24abd38144d8f2ca6a08dbe9a7367662f9388ce6a6 not found: ID does not exist" containerID="38c693208bc6538300642f24abd38144d8f2ca6a08dbe9a7367662f9388ce6a6" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.853443 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38c693208bc6538300642f24abd38144d8f2ca6a08dbe9a7367662f9388ce6a6"} err="failed to get container status \"38c693208bc6538300642f24abd38144d8f2ca6a08dbe9a7367662f9388ce6a6\": rpc error: code = NotFound desc = could not find container \"38c693208bc6538300642f24abd38144d8f2ca6a08dbe9a7367662f9388ce6a6\": container with ID starting with 38c693208bc6538300642f24abd38144d8f2ca6a08dbe9a7367662f9388ce6a6 not found: ID does not exist" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.853463 4771 scope.go:117] "RemoveContainer" containerID="0d070db56d170aa762f1b190c3e84f7efc974e30e582ec7808ec7d8bfb6912bc" Oct 02 10:48:47 crc kubenswrapper[4771]: E1002 10:48:47.853770 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d070db56d170aa762f1b190c3e84f7efc974e30e582ec7808ec7d8bfb6912bc\": container with ID starting with 0d070db56d170aa762f1b190c3e84f7efc974e30e582ec7808ec7d8bfb6912bc not found: ID does not exist" containerID="0d070db56d170aa762f1b190c3e84f7efc974e30e582ec7808ec7d8bfb6912bc" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.853803 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d070db56d170aa762f1b190c3e84f7efc974e30e582ec7808ec7d8bfb6912bc"} err="failed to get container status \"0d070db56d170aa762f1b190c3e84f7efc974e30e582ec7808ec7d8bfb6912bc\": rpc error: code = NotFound desc = could not find container \"0d070db56d170aa762f1b190c3e84f7efc974e30e582ec7808ec7d8bfb6912bc\": container with ID starting with 0d070db56d170aa762f1b190c3e84f7efc974e30e582ec7808ec7d8bfb6912bc not found: ID does not exist" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.853826 4771 scope.go:117] "RemoveContainer" containerID="e74df73d8ce599ec6dc23cc26366c4f678a916f72a21dd0b3ee37e0440649efe" Oct 02 10:48:47 crc kubenswrapper[4771]: E1002 10:48:47.854629 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e74df73d8ce599ec6dc23cc26366c4f678a916f72a21dd0b3ee37e0440649efe\": container with ID starting with e74df73d8ce599ec6dc23cc26366c4f678a916f72a21dd0b3ee37e0440649efe not found: ID does not exist" containerID="e74df73d8ce599ec6dc23cc26366c4f678a916f72a21dd0b3ee37e0440649efe" Oct 02 10:48:47 crc kubenswrapper[4771]: I1002 10:48:47.854751 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e74df73d8ce599ec6dc23cc26366c4f678a916f72a21dd0b3ee37e0440649efe"} err="failed to get container status \"e74df73d8ce599ec6dc23cc26366c4f678a916f72a21dd0b3ee37e0440649efe\": rpc error: code = NotFound desc = could not find container \"e74df73d8ce599ec6dc23cc26366c4f678a916f72a21dd0b3ee37e0440649efe\": container with ID starting with e74df73d8ce599ec6dc23cc26366c4f678a916f72a21dd0b3ee37e0440649efe not found: ID does not exist" Oct 02 10:48:49 crc kubenswrapper[4771]: I1002 10:48:49.695325 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1ece6dc-bff0-436d-a81b-8963a04162e9" path="/var/lib/kubelet/pods/f1ece6dc-bff0-436d-a81b-8963a04162e9/volumes" Oct 02 10:49:12 crc kubenswrapper[4771]: I1002 10:49:12.145982 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:49:12 crc kubenswrapper[4771]: I1002 10:49:12.146583 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:49:12 crc kubenswrapper[4771]: I1002 10:49:12.146641 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 10:49:12 crc kubenswrapper[4771]: I1002 10:49:12.147633 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:49:12 crc kubenswrapper[4771]: I1002 10:49:12.147694 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" gracePeriod=600 Oct 02 10:49:12 crc kubenswrapper[4771]: E1002 10:49:12.816868 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:49:13 crc kubenswrapper[4771]: I1002 10:49:13.017048 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" exitCode=0 Oct 02 10:49:13 crc kubenswrapper[4771]: I1002 10:49:13.017140 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764"} Oct 02 10:49:13 crc kubenswrapper[4771]: I1002 10:49:13.017388 4771 scope.go:117] "RemoveContainer" containerID="194c5b729f6dfb0975af45a7ef28aedac3730958132074ff9a427ab736eddfcb" Oct 02 10:49:13 crc kubenswrapper[4771]: I1002 10:49:13.018171 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:49:13 crc kubenswrapper[4771]: E1002 10:49:13.018469 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:49:27 crc kubenswrapper[4771]: I1002 10:49:27.681753 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:49:27 crc kubenswrapper[4771]: E1002 10:49:27.682580 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:49:41 crc kubenswrapper[4771]: I1002 10:49:41.687457 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:49:41 crc kubenswrapper[4771]: E1002 10:49:41.688749 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:49:54 crc kubenswrapper[4771]: I1002 10:49:54.682580 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:49:54 crc kubenswrapper[4771]: E1002 10:49:54.683478 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:50:05 crc kubenswrapper[4771]: I1002 10:50:05.000809 4771 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-668c9ddd79-9f298" podUID="285035a2-e600-4571-9462-e029a84b3779" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 02 10:50:07 crc kubenswrapper[4771]: I1002 10:50:07.686449 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:50:07 crc kubenswrapper[4771]: E1002 10:50:07.687563 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:50:18 crc kubenswrapper[4771]: I1002 10:50:18.681676 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:50:18 crc kubenswrapper[4771]: E1002 10:50:18.682335 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:50:29 crc kubenswrapper[4771]: I1002 10:50:29.681822 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:50:29 crc kubenswrapper[4771]: E1002 10:50:29.682644 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:50:41 crc kubenswrapper[4771]: I1002 10:50:41.681189 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:50:41 crc kubenswrapper[4771]: E1002 10:50:41.682612 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:50:52 crc kubenswrapper[4771]: I1002 10:50:52.681951 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:50:52 crc kubenswrapper[4771]: E1002 10:50:52.682705 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:51:06 crc kubenswrapper[4771]: I1002 10:51:06.682611 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:51:06 crc kubenswrapper[4771]: E1002 10:51:06.683770 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:51:20 crc kubenswrapper[4771]: I1002 10:51:20.684005 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:51:20 crc kubenswrapper[4771]: E1002 10:51:20.684826 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:51:33 crc kubenswrapper[4771]: I1002 10:51:33.692850 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:51:33 crc kubenswrapper[4771]: E1002 10:51:33.695960 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:51:46 crc kubenswrapper[4771]: I1002 10:51:46.681768 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:51:46 crc kubenswrapper[4771]: E1002 10:51:46.682658 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:51:58 crc kubenswrapper[4771]: I1002 10:51:58.683559 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:51:58 crc kubenswrapper[4771]: E1002 10:51:58.684633 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:52:12 crc kubenswrapper[4771]: I1002 10:52:12.681844 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:52:12 crc kubenswrapper[4771]: E1002 10:52:12.682932 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:52:25 crc kubenswrapper[4771]: I1002 10:52:25.681361 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:52:25 crc kubenswrapper[4771]: E1002 10:52:25.682201 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:52:37 crc kubenswrapper[4771]: I1002 10:52:37.682174 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:52:37 crc kubenswrapper[4771]: E1002 10:52:37.683327 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:52:50 crc kubenswrapper[4771]: I1002 10:52:50.682446 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:52:50 crc kubenswrapper[4771]: E1002 10:52:50.683411 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:53:02 crc kubenswrapper[4771]: I1002 10:53:02.682277 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:53:02 crc kubenswrapper[4771]: E1002 10:53:02.683242 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:53:17 crc kubenswrapper[4771]: I1002 10:53:17.681181 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:53:17 crc kubenswrapper[4771]: E1002 10:53:17.681975 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.646262 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9zs2g"] Oct 02 10:53:31 crc kubenswrapper[4771]: E1002 10:53:31.647229 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869d9a08-35b0-4cf9-97a6-b5939a32ffb4" containerName="extract-utilities" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.647243 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="869d9a08-35b0-4cf9-97a6-b5939a32ffb4" containerName="extract-utilities" Oct 02 10:53:31 crc kubenswrapper[4771]: E1002 10:53:31.647261 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869d9a08-35b0-4cf9-97a6-b5939a32ffb4" containerName="registry-server" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.647268 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="869d9a08-35b0-4cf9-97a6-b5939a32ffb4" containerName="registry-server" Oct 02 10:53:31 crc kubenswrapper[4771]: E1002 10:53:31.647287 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ece6dc-bff0-436d-a81b-8963a04162e9" containerName="registry-server" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.647293 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ece6dc-bff0-436d-a81b-8963a04162e9" containerName="registry-server" Oct 02 10:53:31 crc kubenswrapper[4771]: E1002 10:53:31.647315 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" containerName="extract-utilities" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.647321 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" containerName="extract-utilities" Oct 02 10:53:31 crc kubenswrapper[4771]: E1002 10:53:31.647339 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" containerName="extract-content" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.647345 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" containerName="extract-content" Oct 02 10:53:31 crc kubenswrapper[4771]: E1002 10:53:31.647359 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ece6dc-bff0-436d-a81b-8963a04162e9" containerName="extract-content" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.647364 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ece6dc-bff0-436d-a81b-8963a04162e9" containerName="extract-content" Oct 02 10:53:31 crc kubenswrapper[4771]: E1002 10:53:31.647378 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869d9a08-35b0-4cf9-97a6-b5939a32ffb4" containerName="extract-content" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.647385 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="869d9a08-35b0-4cf9-97a6-b5939a32ffb4" containerName="extract-content" Oct 02 10:53:31 crc kubenswrapper[4771]: E1002 10:53:31.647407 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ece6dc-bff0-436d-a81b-8963a04162e9" containerName="extract-utilities" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.647418 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ece6dc-bff0-436d-a81b-8963a04162e9" containerName="extract-utilities" Oct 02 10:53:31 crc kubenswrapper[4771]: E1002 10:53:31.647451 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" containerName="registry-server" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.647457 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" containerName="registry-server" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.647709 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1ece6dc-bff0-436d-a81b-8963a04162e9" containerName="registry-server" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.647739 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa69f30-04cc-4a8d-848e-fdc9ee42b1d5" containerName="registry-server" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.647754 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="869d9a08-35b0-4cf9-97a6-b5939a32ffb4" containerName="registry-server" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.649701 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.660640 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9zs2g"] Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.690790 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:53:31 crc kubenswrapper[4771]: E1002 10:53:31.691235 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.752390 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6aca7a-b069-494d-999b-f93dc1ed92e0-catalog-content\") pod \"redhat-operators-9zs2g\" (UID: \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\") " pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.752839 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prkfk\" (UniqueName: \"kubernetes.io/projected/5f6aca7a-b069-494d-999b-f93dc1ed92e0-kube-api-access-prkfk\") pod \"redhat-operators-9zs2g\" (UID: \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\") " pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.752920 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6aca7a-b069-494d-999b-f93dc1ed92e0-utilities\") pod \"redhat-operators-9zs2g\" (UID: \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\") " pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.854945 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prkfk\" (UniqueName: \"kubernetes.io/projected/5f6aca7a-b069-494d-999b-f93dc1ed92e0-kube-api-access-prkfk\") pod \"redhat-operators-9zs2g\" (UID: \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\") " pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.855030 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6aca7a-b069-494d-999b-f93dc1ed92e0-utilities\") pod \"redhat-operators-9zs2g\" (UID: \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\") " pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.855090 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6aca7a-b069-494d-999b-f93dc1ed92e0-catalog-content\") pod \"redhat-operators-9zs2g\" (UID: \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\") " pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.855804 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6aca7a-b069-494d-999b-f93dc1ed92e0-utilities\") pod \"redhat-operators-9zs2g\" (UID: \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\") " pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.856061 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6aca7a-b069-494d-999b-f93dc1ed92e0-catalog-content\") pod \"redhat-operators-9zs2g\" (UID: \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\") " pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:53:31 crc kubenswrapper[4771]: I1002 10:53:31.878024 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prkfk\" (UniqueName: \"kubernetes.io/projected/5f6aca7a-b069-494d-999b-f93dc1ed92e0-kube-api-access-prkfk\") pod \"redhat-operators-9zs2g\" (UID: \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\") " pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:53:32 crc kubenswrapper[4771]: I1002 10:53:32.013947 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:53:32 crc kubenswrapper[4771]: I1002 10:53:32.518982 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9zs2g"] Oct 02 10:53:32 crc kubenswrapper[4771]: I1002 10:53:32.969350 4771 generic.go:334] "Generic (PLEG): container finished" podID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerID="2228ee4c85f193eca5903a4179fe4b7bf22a0c4ddab46c1d382199c7e4c1e5ab" exitCode=0 Oct 02 10:53:32 crc kubenswrapper[4771]: I1002 10:53:32.969455 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9zs2g" event={"ID":"5f6aca7a-b069-494d-999b-f93dc1ed92e0","Type":"ContainerDied","Data":"2228ee4c85f193eca5903a4179fe4b7bf22a0c4ddab46c1d382199c7e4c1e5ab"} Oct 02 10:53:32 crc kubenswrapper[4771]: I1002 10:53:32.969770 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9zs2g" event={"ID":"5f6aca7a-b069-494d-999b-f93dc1ed92e0","Type":"ContainerStarted","Data":"1f63bc8f61cd4175210231e58fde78885fa3f9608740270c202e851109d2172c"} Oct 02 10:53:32 crc kubenswrapper[4771]: I1002 10:53:32.972637 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:53:33 crc kubenswrapper[4771]: I1002 10:53:33.981055 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9zs2g" event={"ID":"5f6aca7a-b069-494d-999b-f93dc1ed92e0","Type":"ContainerStarted","Data":"f5fbb396641fce20999316c750b61bac8a3e8e2be80e5b0238766d33fbc4a17f"} Oct 02 10:53:40 crc kubenswrapper[4771]: I1002 10:53:40.041315 4771 generic.go:334] "Generic (PLEG): container finished" podID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerID="f5fbb396641fce20999316c750b61bac8a3e8e2be80e5b0238766d33fbc4a17f" exitCode=0 Oct 02 10:53:40 crc kubenswrapper[4771]: I1002 10:53:40.041444 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9zs2g" event={"ID":"5f6aca7a-b069-494d-999b-f93dc1ed92e0","Type":"ContainerDied","Data":"f5fbb396641fce20999316c750b61bac8a3e8e2be80e5b0238766d33fbc4a17f"} Oct 02 10:53:41 crc kubenswrapper[4771]: I1002 10:53:41.054152 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9zs2g" event={"ID":"5f6aca7a-b069-494d-999b-f93dc1ed92e0","Type":"ContainerStarted","Data":"2cd0d6fcaa8d07fbb1d0ee3c7d7bd8b611db99155ca0fb250dcf7789575013d8"} Oct 02 10:53:41 crc kubenswrapper[4771]: I1002 10:53:41.081534 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9zs2g" podStartSLOduration=2.570275343 podStartE2EDuration="10.081515089s" podCreationTimestamp="2025-10-02 10:53:31 +0000 UTC" firstStartedPulling="2025-10-02 10:53:32.972267397 +0000 UTC m=+4600.619952464" lastFinishedPulling="2025-10-02 10:53:40.483507143 +0000 UTC m=+4608.131192210" observedRunningTime="2025-10-02 10:53:41.070024322 +0000 UTC m=+4608.717709389" watchObservedRunningTime="2025-10-02 10:53:41.081515089 +0000 UTC m=+4608.729200156" Oct 02 10:53:42 crc kubenswrapper[4771]: I1002 10:53:42.015221 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:53:42 crc kubenswrapper[4771]: I1002 10:53:42.015570 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:53:43 crc kubenswrapper[4771]: I1002 10:53:43.066160 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9zs2g" podUID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerName="registry-server" probeResult="failure" output=< Oct 02 10:53:43 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:53:43 crc kubenswrapper[4771]: > Oct 02 10:53:46 crc kubenswrapper[4771]: I1002 10:53:46.682284 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:53:46 crc kubenswrapper[4771]: E1002 10:53:46.683144 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:53:53 crc kubenswrapper[4771]: I1002 10:53:53.065044 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9zs2g" podUID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerName="registry-server" probeResult="failure" output=< Oct 02 10:53:53 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:53:53 crc kubenswrapper[4771]: > Oct 02 10:54:01 crc kubenswrapper[4771]: I1002 10:54:01.682624 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:54:01 crc kubenswrapper[4771]: E1002 10:54:01.683922 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 10:54:03 crc kubenswrapper[4771]: I1002 10:54:03.066380 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9zs2g" podUID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerName="registry-server" probeResult="failure" output=< Oct 02 10:54:03 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:54:03 crc kubenswrapper[4771]: > Oct 02 10:54:12 crc kubenswrapper[4771]: I1002 10:54:12.087756 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:54:12 crc kubenswrapper[4771]: I1002 10:54:12.150753 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:54:12 crc kubenswrapper[4771]: I1002 10:54:12.349846 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9zs2g"] Oct 02 10:54:13 crc kubenswrapper[4771]: I1002 10:54:13.478920 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9zs2g" podUID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerName="registry-server" containerID="cri-o://2cd0d6fcaa8d07fbb1d0ee3c7d7bd8b611db99155ca0fb250dcf7789575013d8" gracePeriod=2 Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.096187 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.245509 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6aca7a-b069-494d-999b-f93dc1ed92e0-utilities\") pod \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\" (UID: \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\") " Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.245682 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prkfk\" (UniqueName: \"kubernetes.io/projected/5f6aca7a-b069-494d-999b-f93dc1ed92e0-kube-api-access-prkfk\") pod \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\" (UID: \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\") " Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.245742 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6aca7a-b069-494d-999b-f93dc1ed92e0-catalog-content\") pod \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\" (UID: \"5f6aca7a-b069-494d-999b-f93dc1ed92e0\") " Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.247376 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f6aca7a-b069-494d-999b-f93dc1ed92e0-utilities" (OuterVolumeSpecName: "utilities") pod "5f6aca7a-b069-494d-999b-f93dc1ed92e0" (UID: "5f6aca7a-b069-494d-999b-f93dc1ed92e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.252929 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f6aca7a-b069-494d-999b-f93dc1ed92e0-kube-api-access-prkfk" (OuterVolumeSpecName: "kube-api-access-prkfk") pod "5f6aca7a-b069-494d-999b-f93dc1ed92e0" (UID: "5f6aca7a-b069-494d-999b-f93dc1ed92e0"). InnerVolumeSpecName "kube-api-access-prkfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.319727 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f6aca7a-b069-494d-999b-f93dc1ed92e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f6aca7a-b069-494d-999b-f93dc1ed92e0" (UID: "5f6aca7a-b069-494d-999b-f93dc1ed92e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.350838 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6aca7a-b069-494d-999b-f93dc1ed92e0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.350924 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prkfk\" (UniqueName: \"kubernetes.io/projected/5f6aca7a-b069-494d-999b-f93dc1ed92e0-kube-api-access-prkfk\") on node \"crc\" DevicePath \"\"" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.350937 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6aca7a-b069-494d-999b-f93dc1ed92e0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.494206 4771 generic.go:334] "Generic (PLEG): container finished" podID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerID="2cd0d6fcaa8d07fbb1d0ee3c7d7bd8b611db99155ca0fb250dcf7789575013d8" exitCode=0 Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.494259 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9zs2g" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.494257 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9zs2g" event={"ID":"5f6aca7a-b069-494d-999b-f93dc1ed92e0","Type":"ContainerDied","Data":"2cd0d6fcaa8d07fbb1d0ee3c7d7bd8b611db99155ca0fb250dcf7789575013d8"} Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.494375 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9zs2g" event={"ID":"5f6aca7a-b069-494d-999b-f93dc1ed92e0","Type":"ContainerDied","Data":"1f63bc8f61cd4175210231e58fde78885fa3f9608740270c202e851109d2172c"} Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.494403 4771 scope.go:117] "RemoveContainer" containerID="2cd0d6fcaa8d07fbb1d0ee3c7d7bd8b611db99155ca0fb250dcf7789575013d8" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.519557 4771 scope.go:117] "RemoveContainer" containerID="f5fbb396641fce20999316c750b61bac8a3e8e2be80e5b0238766d33fbc4a17f" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.534397 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9zs2g"] Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.546356 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9zs2g"] Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.571648 4771 scope.go:117] "RemoveContainer" containerID="2228ee4c85f193eca5903a4179fe4b7bf22a0c4ddab46c1d382199c7e4c1e5ab" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.597825 4771 scope.go:117] "RemoveContainer" containerID="2cd0d6fcaa8d07fbb1d0ee3c7d7bd8b611db99155ca0fb250dcf7789575013d8" Oct 02 10:54:14 crc kubenswrapper[4771]: E1002 10:54:14.598343 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cd0d6fcaa8d07fbb1d0ee3c7d7bd8b611db99155ca0fb250dcf7789575013d8\": container with ID starting with 2cd0d6fcaa8d07fbb1d0ee3c7d7bd8b611db99155ca0fb250dcf7789575013d8 not found: ID does not exist" containerID="2cd0d6fcaa8d07fbb1d0ee3c7d7bd8b611db99155ca0fb250dcf7789575013d8" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.598389 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cd0d6fcaa8d07fbb1d0ee3c7d7bd8b611db99155ca0fb250dcf7789575013d8"} err="failed to get container status \"2cd0d6fcaa8d07fbb1d0ee3c7d7bd8b611db99155ca0fb250dcf7789575013d8\": rpc error: code = NotFound desc = could not find container \"2cd0d6fcaa8d07fbb1d0ee3c7d7bd8b611db99155ca0fb250dcf7789575013d8\": container with ID starting with 2cd0d6fcaa8d07fbb1d0ee3c7d7bd8b611db99155ca0fb250dcf7789575013d8 not found: ID does not exist" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.598415 4771 scope.go:117] "RemoveContainer" containerID="f5fbb396641fce20999316c750b61bac8a3e8e2be80e5b0238766d33fbc4a17f" Oct 02 10:54:14 crc kubenswrapper[4771]: E1002 10:54:14.598712 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5fbb396641fce20999316c750b61bac8a3e8e2be80e5b0238766d33fbc4a17f\": container with ID starting with f5fbb396641fce20999316c750b61bac8a3e8e2be80e5b0238766d33fbc4a17f not found: ID does not exist" containerID="f5fbb396641fce20999316c750b61bac8a3e8e2be80e5b0238766d33fbc4a17f" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.598741 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5fbb396641fce20999316c750b61bac8a3e8e2be80e5b0238766d33fbc4a17f"} err="failed to get container status \"f5fbb396641fce20999316c750b61bac8a3e8e2be80e5b0238766d33fbc4a17f\": rpc error: code = NotFound desc = could not find container \"f5fbb396641fce20999316c750b61bac8a3e8e2be80e5b0238766d33fbc4a17f\": container with ID starting with f5fbb396641fce20999316c750b61bac8a3e8e2be80e5b0238766d33fbc4a17f not found: ID does not exist" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.598759 4771 scope.go:117] "RemoveContainer" containerID="2228ee4c85f193eca5903a4179fe4b7bf22a0c4ddab46c1d382199c7e4c1e5ab" Oct 02 10:54:14 crc kubenswrapper[4771]: E1002 10:54:14.599227 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2228ee4c85f193eca5903a4179fe4b7bf22a0c4ddab46c1d382199c7e4c1e5ab\": container with ID starting with 2228ee4c85f193eca5903a4179fe4b7bf22a0c4ddab46c1d382199c7e4c1e5ab not found: ID does not exist" containerID="2228ee4c85f193eca5903a4179fe4b7bf22a0c4ddab46c1d382199c7e4c1e5ab" Oct 02 10:54:14 crc kubenswrapper[4771]: I1002 10:54:14.599251 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2228ee4c85f193eca5903a4179fe4b7bf22a0c4ddab46c1d382199c7e4c1e5ab"} err="failed to get container status \"2228ee4c85f193eca5903a4179fe4b7bf22a0c4ddab46c1d382199c7e4c1e5ab\": rpc error: code = NotFound desc = could not find container \"2228ee4c85f193eca5903a4179fe4b7bf22a0c4ddab46c1d382199c7e4c1e5ab\": container with ID starting with 2228ee4c85f193eca5903a4179fe4b7bf22a0c4ddab46c1d382199c7e4c1e5ab not found: ID does not exist" Oct 02 10:54:15 crc kubenswrapper[4771]: I1002 10:54:15.681971 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:54:15 crc kubenswrapper[4771]: I1002 10:54:15.699329 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" path="/var/lib/kubelet/pods/5f6aca7a-b069-494d-999b-f93dc1ed92e0/volumes" Oct 02 10:54:16 crc kubenswrapper[4771]: I1002 10:54:16.519452 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"3879c77192f81ec753429c0f3f323cb067948912c2feb70ad2e6e8689e44370a"} Oct 02 10:56:42 crc kubenswrapper[4771]: I1002 10:56:42.145967 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:56:42 crc kubenswrapper[4771]: I1002 10:56:42.147031 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:57:12 crc kubenswrapper[4771]: E1002 10:57:12.119298 4771 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.53:53940->38.102.83.53:41045: write tcp 38.102.83.53:53940->38.102.83.53:41045: write: broken pipe Oct 02 10:57:12 crc kubenswrapper[4771]: I1002 10:57:12.146074 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:57:12 crc kubenswrapper[4771]: I1002 10:57:12.146231 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:57:42 crc kubenswrapper[4771]: I1002 10:57:42.146448 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:57:42 crc kubenswrapper[4771]: I1002 10:57:42.146955 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:57:42 crc kubenswrapper[4771]: I1002 10:57:42.146999 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 10:57:42 crc kubenswrapper[4771]: I1002 10:57:42.149159 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3879c77192f81ec753429c0f3f323cb067948912c2feb70ad2e6e8689e44370a"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:57:42 crc kubenswrapper[4771]: I1002 10:57:42.149253 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://3879c77192f81ec753429c0f3f323cb067948912c2feb70ad2e6e8689e44370a" gracePeriod=600 Oct 02 10:57:42 crc kubenswrapper[4771]: I1002 10:57:42.910523 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="3879c77192f81ec753429c0f3f323cb067948912c2feb70ad2e6e8689e44370a" exitCode=0 Oct 02 10:57:42 crc kubenswrapper[4771]: I1002 10:57:42.910601 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"3879c77192f81ec753429c0f3f323cb067948912c2feb70ad2e6e8689e44370a"} Oct 02 10:57:42 crc kubenswrapper[4771]: I1002 10:57:42.910885 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b"} Oct 02 10:57:42 crc kubenswrapper[4771]: I1002 10:57:42.910909 4771 scope.go:117] "RemoveContainer" containerID="2821fa7013a3121c0f3d6ab4df2eb8db93ed3ae313f6974b22e90bf2effb8764" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.563771 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cq4j6"] Oct 02 10:58:31 crc kubenswrapper[4771]: E1002 10:58:31.564984 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerName="extract-content" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.565000 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerName="extract-content" Oct 02 10:58:31 crc kubenswrapper[4771]: E1002 10:58:31.565018 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerName="registry-server" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.565024 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerName="registry-server" Oct 02 10:58:31 crc kubenswrapper[4771]: E1002 10:58:31.565066 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerName="extract-utilities" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.565073 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerName="extract-utilities" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.565328 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f6aca7a-b069-494d-999b-f93dc1ed92e0" containerName="registry-server" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.567161 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.578236 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cq4j6"] Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.581150 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b729c0-e950-48d8-b998-1676e941a9a1-catalog-content\") pod \"redhat-marketplace-cq4j6\" (UID: \"f3b729c0-e950-48d8-b998-1676e941a9a1\") " pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.581222 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b729c0-e950-48d8-b998-1676e941a9a1-utilities\") pod \"redhat-marketplace-cq4j6\" (UID: \"f3b729c0-e950-48d8-b998-1676e941a9a1\") " pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.581298 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvs2x\" (UniqueName: \"kubernetes.io/projected/f3b729c0-e950-48d8-b998-1676e941a9a1-kube-api-access-mvs2x\") pod \"redhat-marketplace-cq4j6\" (UID: \"f3b729c0-e950-48d8-b998-1676e941a9a1\") " pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.684186 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b729c0-e950-48d8-b998-1676e941a9a1-catalog-content\") pod \"redhat-marketplace-cq4j6\" (UID: \"f3b729c0-e950-48d8-b998-1676e941a9a1\") " pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.684270 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b729c0-e950-48d8-b998-1676e941a9a1-utilities\") pod \"redhat-marketplace-cq4j6\" (UID: \"f3b729c0-e950-48d8-b998-1676e941a9a1\") " pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.684352 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvs2x\" (UniqueName: \"kubernetes.io/projected/f3b729c0-e950-48d8-b998-1676e941a9a1-kube-api-access-mvs2x\") pod \"redhat-marketplace-cq4j6\" (UID: \"f3b729c0-e950-48d8-b998-1676e941a9a1\") " pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.685019 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b729c0-e950-48d8-b998-1676e941a9a1-catalog-content\") pod \"redhat-marketplace-cq4j6\" (UID: \"f3b729c0-e950-48d8-b998-1676e941a9a1\") " pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.685403 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b729c0-e950-48d8-b998-1676e941a9a1-utilities\") pod \"redhat-marketplace-cq4j6\" (UID: \"f3b729c0-e950-48d8-b998-1676e941a9a1\") " pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.711080 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvs2x\" (UniqueName: \"kubernetes.io/projected/f3b729c0-e950-48d8-b998-1676e941a9a1-kube-api-access-mvs2x\") pod \"redhat-marketplace-cq4j6\" (UID: \"f3b729c0-e950-48d8-b998-1676e941a9a1\") " pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:31 crc kubenswrapper[4771]: I1002 10:58:31.891084 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:32 crc kubenswrapper[4771]: I1002 10:58:32.452406 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cq4j6"] Oct 02 10:58:32 crc kubenswrapper[4771]: I1002 10:58:32.552895 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cq4j6" event={"ID":"f3b729c0-e950-48d8-b998-1676e941a9a1","Type":"ContainerStarted","Data":"d9e1fe61310ec2e64659de574a702aeee2483adcc4a874bb1bdd0db9f898d2e6"} Oct 02 10:58:33 crc kubenswrapper[4771]: I1002 10:58:33.565606 4771 generic.go:334] "Generic (PLEG): container finished" podID="f3b729c0-e950-48d8-b998-1676e941a9a1" containerID="e89b5d05bb9e1c594b7924994514064b35a079810025c26187be5c019040cef7" exitCode=0 Oct 02 10:58:33 crc kubenswrapper[4771]: I1002 10:58:33.565694 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cq4j6" event={"ID":"f3b729c0-e950-48d8-b998-1676e941a9a1","Type":"ContainerDied","Data":"e89b5d05bb9e1c594b7924994514064b35a079810025c26187be5c019040cef7"} Oct 02 10:58:33 crc kubenswrapper[4771]: I1002 10:58:33.568876 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:58:34 crc kubenswrapper[4771]: I1002 10:58:34.578860 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cq4j6" event={"ID":"f3b729c0-e950-48d8-b998-1676e941a9a1","Type":"ContainerStarted","Data":"ea1c22bc197bab815cdf208aeaf85ee5e7b002ea9d60af8a05bb1fa2eb7431fa"} Oct 02 10:58:35 crc kubenswrapper[4771]: I1002 10:58:35.590847 4771 generic.go:334] "Generic (PLEG): container finished" podID="f3b729c0-e950-48d8-b998-1676e941a9a1" containerID="ea1c22bc197bab815cdf208aeaf85ee5e7b002ea9d60af8a05bb1fa2eb7431fa" exitCode=0 Oct 02 10:58:35 crc kubenswrapper[4771]: I1002 10:58:35.590892 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cq4j6" event={"ID":"f3b729c0-e950-48d8-b998-1676e941a9a1","Type":"ContainerDied","Data":"ea1c22bc197bab815cdf208aeaf85ee5e7b002ea9d60af8a05bb1fa2eb7431fa"} Oct 02 10:58:36 crc kubenswrapper[4771]: I1002 10:58:36.617568 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cq4j6" event={"ID":"f3b729c0-e950-48d8-b998-1676e941a9a1","Type":"ContainerStarted","Data":"747eb23d70f9ae4ab4a01e7397f8e75fb131e240cdd7982736f52e6e1f0f435c"} Oct 02 10:58:36 crc kubenswrapper[4771]: I1002 10:58:36.649658 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cq4j6" podStartSLOduration=3.11964339 podStartE2EDuration="5.649633071s" podCreationTimestamp="2025-10-02 10:58:31 +0000 UTC" firstStartedPulling="2025-10-02 10:58:33.568627643 +0000 UTC m=+4901.216312710" lastFinishedPulling="2025-10-02 10:58:36.098617314 +0000 UTC m=+4903.746302391" observedRunningTime="2025-10-02 10:58:36.633837765 +0000 UTC m=+4904.281522852" watchObservedRunningTime="2025-10-02 10:58:36.649633071 +0000 UTC m=+4904.297318148" Oct 02 10:58:41 crc kubenswrapper[4771]: I1002 10:58:41.891631 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:41 crc kubenswrapper[4771]: I1002 10:58:41.892242 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:41 crc kubenswrapper[4771]: I1002 10:58:41.948961 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:42 crc kubenswrapper[4771]: I1002 10:58:42.729554 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:42 crc kubenswrapper[4771]: I1002 10:58:42.778571 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cq4j6"] Oct 02 10:58:44 crc kubenswrapper[4771]: I1002 10:58:44.717051 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cq4j6" podUID="f3b729c0-e950-48d8-b998-1676e941a9a1" containerName="registry-server" containerID="cri-o://747eb23d70f9ae4ab4a01e7397f8e75fb131e240cdd7982736f52e6e1f0f435c" gracePeriod=2 Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.253082 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.426024 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b729c0-e950-48d8-b998-1676e941a9a1-utilities\") pod \"f3b729c0-e950-48d8-b998-1676e941a9a1\" (UID: \"f3b729c0-e950-48d8-b998-1676e941a9a1\") " Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.426192 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvs2x\" (UniqueName: \"kubernetes.io/projected/f3b729c0-e950-48d8-b998-1676e941a9a1-kube-api-access-mvs2x\") pod \"f3b729c0-e950-48d8-b998-1676e941a9a1\" (UID: \"f3b729c0-e950-48d8-b998-1676e941a9a1\") " Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.426377 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b729c0-e950-48d8-b998-1676e941a9a1-catalog-content\") pod \"f3b729c0-e950-48d8-b998-1676e941a9a1\" (UID: \"f3b729c0-e950-48d8-b998-1676e941a9a1\") " Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.427225 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3b729c0-e950-48d8-b998-1676e941a9a1-utilities" (OuterVolumeSpecName: "utilities") pod "f3b729c0-e950-48d8-b998-1676e941a9a1" (UID: "f3b729c0-e950-48d8-b998-1676e941a9a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.432466 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3b729c0-e950-48d8-b998-1676e941a9a1-kube-api-access-mvs2x" (OuterVolumeSpecName: "kube-api-access-mvs2x") pod "f3b729c0-e950-48d8-b998-1676e941a9a1" (UID: "f3b729c0-e950-48d8-b998-1676e941a9a1"). InnerVolumeSpecName "kube-api-access-mvs2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.446264 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3b729c0-e950-48d8-b998-1676e941a9a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3b729c0-e950-48d8-b998-1676e941a9a1" (UID: "f3b729c0-e950-48d8-b998-1676e941a9a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.529463 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b729c0-e950-48d8-b998-1676e941a9a1-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.529503 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvs2x\" (UniqueName: \"kubernetes.io/projected/f3b729c0-e950-48d8-b998-1676e941a9a1-kube-api-access-mvs2x\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.529519 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b729c0-e950-48d8-b998-1676e941a9a1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.728543 4771 generic.go:334] "Generic (PLEG): container finished" podID="f3b729c0-e950-48d8-b998-1676e941a9a1" containerID="747eb23d70f9ae4ab4a01e7397f8e75fb131e240cdd7982736f52e6e1f0f435c" exitCode=0 Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.728583 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cq4j6" event={"ID":"f3b729c0-e950-48d8-b998-1676e941a9a1","Type":"ContainerDied","Data":"747eb23d70f9ae4ab4a01e7397f8e75fb131e240cdd7982736f52e6e1f0f435c"} Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.728610 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cq4j6" event={"ID":"f3b729c0-e950-48d8-b998-1676e941a9a1","Type":"ContainerDied","Data":"d9e1fe61310ec2e64659de574a702aeee2483adcc4a874bb1bdd0db9f898d2e6"} Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.728627 4771 scope.go:117] "RemoveContainer" containerID="747eb23d70f9ae4ab4a01e7397f8e75fb131e240cdd7982736f52e6e1f0f435c" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.728700 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cq4j6" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.756709 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cq4j6"] Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.763881 4771 scope.go:117] "RemoveContainer" containerID="ea1c22bc197bab815cdf208aeaf85ee5e7b002ea9d60af8a05bb1fa2eb7431fa" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.768480 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cq4j6"] Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.796565 4771 scope.go:117] "RemoveContainer" containerID="e89b5d05bb9e1c594b7924994514064b35a079810025c26187be5c019040cef7" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.844955 4771 scope.go:117] "RemoveContainer" containerID="747eb23d70f9ae4ab4a01e7397f8e75fb131e240cdd7982736f52e6e1f0f435c" Oct 02 10:58:45 crc kubenswrapper[4771]: E1002 10:58:45.845799 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"747eb23d70f9ae4ab4a01e7397f8e75fb131e240cdd7982736f52e6e1f0f435c\": container with ID starting with 747eb23d70f9ae4ab4a01e7397f8e75fb131e240cdd7982736f52e6e1f0f435c not found: ID does not exist" containerID="747eb23d70f9ae4ab4a01e7397f8e75fb131e240cdd7982736f52e6e1f0f435c" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.845850 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"747eb23d70f9ae4ab4a01e7397f8e75fb131e240cdd7982736f52e6e1f0f435c"} err="failed to get container status \"747eb23d70f9ae4ab4a01e7397f8e75fb131e240cdd7982736f52e6e1f0f435c\": rpc error: code = NotFound desc = could not find container \"747eb23d70f9ae4ab4a01e7397f8e75fb131e240cdd7982736f52e6e1f0f435c\": container with ID starting with 747eb23d70f9ae4ab4a01e7397f8e75fb131e240cdd7982736f52e6e1f0f435c not found: ID does not exist" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.845880 4771 scope.go:117] "RemoveContainer" containerID="ea1c22bc197bab815cdf208aeaf85ee5e7b002ea9d60af8a05bb1fa2eb7431fa" Oct 02 10:58:45 crc kubenswrapper[4771]: E1002 10:58:45.846480 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea1c22bc197bab815cdf208aeaf85ee5e7b002ea9d60af8a05bb1fa2eb7431fa\": container with ID starting with ea1c22bc197bab815cdf208aeaf85ee5e7b002ea9d60af8a05bb1fa2eb7431fa not found: ID does not exist" containerID="ea1c22bc197bab815cdf208aeaf85ee5e7b002ea9d60af8a05bb1fa2eb7431fa" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.846516 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea1c22bc197bab815cdf208aeaf85ee5e7b002ea9d60af8a05bb1fa2eb7431fa"} err="failed to get container status \"ea1c22bc197bab815cdf208aeaf85ee5e7b002ea9d60af8a05bb1fa2eb7431fa\": rpc error: code = NotFound desc = could not find container \"ea1c22bc197bab815cdf208aeaf85ee5e7b002ea9d60af8a05bb1fa2eb7431fa\": container with ID starting with ea1c22bc197bab815cdf208aeaf85ee5e7b002ea9d60af8a05bb1fa2eb7431fa not found: ID does not exist" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.846536 4771 scope.go:117] "RemoveContainer" containerID="e89b5d05bb9e1c594b7924994514064b35a079810025c26187be5c019040cef7" Oct 02 10:58:45 crc kubenswrapper[4771]: E1002 10:58:45.846744 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e89b5d05bb9e1c594b7924994514064b35a079810025c26187be5c019040cef7\": container with ID starting with e89b5d05bb9e1c594b7924994514064b35a079810025c26187be5c019040cef7 not found: ID does not exist" containerID="e89b5d05bb9e1c594b7924994514064b35a079810025c26187be5c019040cef7" Oct 02 10:58:45 crc kubenswrapper[4771]: I1002 10:58:45.846769 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e89b5d05bb9e1c594b7924994514064b35a079810025c26187be5c019040cef7"} err="failed to get container status \"e89b5d05bb9e1c594b7924994514064b35a079810025c26187be5c019040cef7\": rpc error: code = NotFound desc = could not find container \"e89b5d05bb9e1c594b7924994514064b35a079810025c26187be5c019040cef7\": container with ID starting with e89b5d05bb9e1c594b7924994514064b35a079810025c26187be5c019040cef7 not found: ID does not exist" Oct 02 10:58:47 crc kubenswrapper[4771]: I1002 10:58:47.697401 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3b729c0-e950-48d8-b998-1676e941a9a1" path="/var/lib/kubelet/pods/f3b729c0-e950-48d8-b998-1676e941a9a1/volumes" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.726108 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f24sn"] Oct 02 10:58:48 crc kubenswrapper[4771]: E1002 10:58:48.726960 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b729c0-e950-48d8-b998-1676e941a9a1" containerName="extract-content" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.726973 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b729c0-e950-48d8-b998-1676e941a9a1" containerName="extract-content" Oct 02 10:58:48 crc kubenswrapper[4771]: E1002 10:58:48.727003 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b729c0-e950-48d8-b998-1676e941a9a1" containerName="extract-utilities" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.727010 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b729c0-e950-48d8-b998-1676e941a9a1" containerName="extract-utilities" Oct 02 10:58:48 crc kubenswrapper[4771]: E1002 10:58:48.727032 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b729c0-e950-48d8-b998-1676e941a9a1" containerName="registry-server" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.727039 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b729c0-e950-48d8-b998-1676e941a9a1" containerName="registry-server" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.727331 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3b729c0-e950-48d8-b998-1676e941a9a1" containerName="registry-server" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.729086 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.749813 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f24sn"] Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.808912 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn4jz\" (UniqueName: \"kubernetes.io/projected/a4099631-fe1c-4047-bd5c-c46ab0b808c3-kube-api-access-qn4jz\") pod \"community-operators-f24sn\" (UID: \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\") " pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.809009 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4099631-fe1c-4047-bd5c-c46ab0b808c3-utilities\") pod \"community-operators-f24sn\" (UID: \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\") " pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.809300 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4099631-fe1c-4047-bd5c-c46ab0b808c3-catalog-content\") pod \"community-operators-f24sn\" (UID: \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\") " pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.912169 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4099631-fe1c-4047-bd5c-c46ab0b808c3-catalog-content\") pod \"community-operators-f24sn\" (UID: \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\") " pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.912606 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn4jz\" (UniqueName: \"kubernetes.io/projected/a4099631-fe1c-4047-bd5c-c46ab0b808c3-kube-api-access-qn4jz\") pod \"community-operators-f24sn\" (UID: \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\") " pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.912755 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4099631-fe1c-4047-bd5c-c46ab0b808c3-utilities\") pod \"community-operators-f24sn\" (UID: \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\") " pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.913026 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4099631-fe1c-4047-bd5c-c46ab0b808c3-catalog-content\") pod \"community-operators-f24sn\" (UID: \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\") " pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.913066 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4099631-fe1c-4047-bd5c-c46ab0b808c3-utilities\") pod \"community-operators-f24sn\" (UID: \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\") " pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:58:48 crc kubenswrapper[4771]: I1002 10:58:48.945920 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn4jz\" (UniqueName: \"kubernetes.io/projected/a4099631-fe1c-4047-bd5c-c46ab0b808c3-kube-api-access-qn4jz\") pod \"community-operators-f24sn\" (UID: \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\") " pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:58:49 crc kubenswrapper[4771]: I1002 10:58:49.059039 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:58:49 crc kubenswrapper[4771]: I1002 10:58:49.629558 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f24sn"] Oct 02 10:58:49 crc kubenswrapper[4771]: I1002 10:58:49.776249 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f24sn" event={"ID":"a4099631-fe1c-4047-bd5c-c46ab0b808c3","Type":"ContainerStarted","Data":"4dfc0a535f210421409deece9654aa16f37bf8bb4b162ad861347afb22e20b27"} Oct 02 10:58:50 crc kubenswrapper[4771]: I1002 10:58:50.790748 4771 generic.go:334] "Generic (PLEG): container finished" podID="a4099631-fe1c-4047-bd5c-c46ab0b808c3" containerID="55bcd7d2fac02922cb300b393a92ae8a70e839a9d1278c6e32288d4afabcdf81" exitCode=0 Oct 02 10:58:50 crc kubenswrapper[4771]: I1002 10:58:50.790818 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f24sn" event={"ID":"a4099631-fe1c-4047-bd5c-c46ab0b808c3","Type":"ContainerDied","Data":"55bcd7d2fac02922cb300b393a92ae8a70e839a9d1278c6e32288d4afabcdf81"} Oct 02 10:58:52 crc kubenswrapper[4771]: I1002 10:58:52.820351 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f24sn" event={"ID":"a4099631-fe1c-4047-bd5c-c46ab0b808c3","Type":"ContainerStarted","Data":"fd74ef67ecdd38afd00a3e7259fd44fef333a68ad3bf16951d37731e4e68bd26"} Oct 02 10:58:55 crc kubenswrapper[4771]: I1002 10:58:55.854605 4771 generic.go:334] "Generic (PLEG): container finished" podID="a4099631-fe1c-4047-bd5c-c46ab0b808c3" containerID="fd74ef67ecdd38afd00a3e7259fd44fef333a68ad3bf16951d37731e4e68bd26" exitCode=0 Oct 02 10:58:55 crc kubenswrapper[4771]: I1002 10:58:55.854773 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f24sn" event={"ID":"a4099631-fe1c-4047-bd5c-c46ab0b808c3","Type":"ContainerDied","Data":"fd74ef67ecdd38afd00a3e7259fd44fef333a68ad3bf16951d37731e4e68bd26"} Oct 02 10:58:56 crc kubenswrapper[4771]: I1002 10:58:56.874930 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f24sn" event={"ID":"a4099631-fe1c-4047-bd5c-c46ab0b808c3","Type":"ContainerStarted","Data":"81a296cd4ef857dfba5b164537ad5f96e8499d2000ff15f91787b1fd4a8e4d7b"} Oct 02 10:58:56 crc kubenswrapper[4771]: I1002 10:58:56.929610 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f24sn" podStartSLOduration=3.346081857 podStartE2EDuration="8.929583011s" podCreationTimestamp="2025-10-02 10:58:48 +0000 UTC" firstStartedPulling="2025-10-02 10:58:50.794902965 +0000 UTC m=+4918.442588032" lastFinishedPulling="2025-10-02 10:58:56.378404119 +0000 UTC m=+4924.026089186" observedRunningTime="2025-10-02 10:58:56.913003474 +0000 UTC m=+4924.560688581" watchObservedRunningTime="2025-10-02 10:58:56.929583011 +0000 UTC m=+4924.577268078" Oct 02 10:58:59 crc kubenswrapper[4771]: I1002 10:58:59.060499 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:58:59 crc kubenswrapper[4771]: I1002 10:58:59.061540 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:59:00 crc kubenswrapper[4771]: I1002 10:59:00.119972 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-f24sn" podUID="a4099631-fe1c-4047-bd5c-c46ab0b808c3" containerName="registry-server" probeResult="failure" output=< Oct 02 10:59:00 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 10:59:00 crc kubenswrapper[4771]: > Oct 02 10:59:09 crc kubenswrapper[4771]: I1002 10:59:09.119337 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:59:09 crc kubenswrapper[4771]: I1002 10:59:09.179602 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:59:09 crc kubenswrapper[4771]: I1002 10:59:09.356004 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f24sn"] Oct 02 10:59:11 crc kubenswrapper[4771]: I1002 10:59:11.042579 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f24sn" podUID="a4099631-fe1c-4047-bd5c-c46ab0b808c3" containerName="registry-server" containerID="cri-o://81a296cd4ef857dfba5b164537ad5f96e8499d2000ff15f91787b1fd4a8e4d7b" gracePeriod=2 Oct 02 10:59:11 crc kubenswrapper[4771]: I1002 10:59:11.640611 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:59:11 crc kubenswrapper[4771]: I1002 10:59:11.723096 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4099631-fe1c-4047-bd5c-c46ab0b808c3-catalog-content\") pod \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\" (UID: \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\") " Oct 02 10:59:11 crc kubenswrapper[4771]: I1002 10:59:11.723265 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn4jz\" (UniqueName: \"kubernetes.io/projected/a4099631-fe1c-4047-bd5c-c46ab0b808c3-kube-api-access-qn4jz\") pod \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\" (UID: \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\") " Oct 02 10:59:11 crc kubenswrapper[4771]: I1002 10:59:11.723343 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4099631-fe1c-4047-bd5c-c46ab0b808c3-utilities\") pod \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\" (UID: \"a4099631-fe1c-4047-bd5c-c46ab0b808c3\") " Oct 02 10:59:11 crc kubenswrapper[4771]: I1002 10:59:11.725275 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4099631-fe1c-4047-bd5c-c46ab0b808c3-utilities" (OuterVolumeSpecName: "utilities") pod "a4099631-fe1c-4047-bd5c-c46ab0b808c3" (UID: "a4099631-fe1c-4047-bd5c-c46ab0b808c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:59:11 crc kubenswrapper[4771]: I1002 10:59:11.738724 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4099631-fe1c-4047-bd5c-c46ab0b808c3-kube-api-access-qn4jz" (OuterVolumeSpecName: "kube-api-access-qn4jz") pod "a4099631-fe1c-4047-bd5c-c46ab0b808c3" (UID: "a4099631-fe1c-4047-bd5c-c46ab0b808c3"). InnerVolumeSpecName "kube-api-access-qn4jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:59:11 crc kubenswrapper[4771]: I1002 10:59:11.826515 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn4jz\" (UniqueName: \"kubernetes.io/projected/a4099631-fe1c-4047-bd5c-c46ab0b808c3-kube-api-access-qn4jz\") on node \"crc\" DevicePath \"\"" Oct 02 10:59:11 crc kubenswrapper[4771]: I1002 10:59:11.826551 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4099631-fe1c-4047-bd5c-c46ab0b808c3-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:59:11 crc kubenswrapper[4771]: I1002 10:59:11.840894 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4099631-fe1c-4047-bd5c-c46ab0b808c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4099631-fe1c-4047-bd5c-c46ab0b808c3" (UID: "a4099631-fe1c-4047-bd5c-c46ab0b808c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:59:11 crc kubenswrapper[4771]: I1002 10:59:11.928618 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4099631-fe1c-4047-bd5c-c46ab0b808c3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.055728 4771 generic.go:334] "Generic (PLEG): container finished" podID="a4099631-fe1c-4047-bd5c-c46ab0b808c3" containerID="81a296cd4ef857dfba5b164537ad5f96e8499d2000ff15f91787b1fd4a8e4d7b" exitCode=0 Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.055776 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f24sn" event={"ID":"a4099631-fe1c-4047-bd5c-c46ab0b808c3","Type":"ContainerDied","Data":"81a296cd4ef857dfba5b164537ad5f96e8499d2000ff15f91787b1fd4a8e4d7b"} Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.055808 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f24sn" event={"ID":"a4099631-fe1c-4047-bd5c-c46ab0b808c3","Type":"ContainerDied","Data":"4dfc0a535f210421409deece9654aa16f37bf8bb4b162ad861347afb22e20b27"} Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.055825 4771 scope.go:117] "RemoveContainer" containerID="81a296cd4ef857dfba5b164537ad5f96e8499d2000ff15f91787b1fd4a8e4d7b" Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.055888 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f24sn" Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.083554 4771 scope.go:117] "RemoveContainer" containerID="fd74ef67ecdd38afd00a3e7259fd44fef333a68ad3bf16951d37731e4e68bd26" Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.099832 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f24sn"] Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.111817 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f24sn"] Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.113827 4771 scope.go:117] "RemoveContainer" containerID="55bcd7d2fac02922cb300b393a92ae8a70e839a9d1278c6e32288d4afabcdf81" Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.175773 4771 scope.go:117] "RemoveContainer" containerID="81a296cd4ef857dfba5b164537ad5f96e8499d2000ff15f91787b1fd4a8e4d7b" Oct 02 10:59:12 crc kubenswrapper[4771]: E1002 10:59:12.176469 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81a296cd4ef857dfba5b164537ad5f96e8499d2000ff15f91787b1fd4a8e4d7b\": container with ID starting with 81a296cd4ef857dfba5b164537ad5f96e8499d2000ff15f91787b1fd4a8e4d7b not found: ID does not exist" containerID="81a296cd4ef857dfba5b164537ad5f96e8499d2000ff15f91787b1fd4a8e4d7b" Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.176506 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81a296cd4ef857dfba5b164537ad5f96e8499d2000ff15f91787b1fd4a8e4d7b"} err="failed to get container status \"81a296cd4ef857dfba5b164537ad5f96e8499d2000ff15f91787b1fd4a8e4d7b\": rpc error: code = NotFound desc = could not find container \"81a296cd4ef857dfba5b164537ad5f96e8499d2000ff15f91787b1fd4a8e4d7b\": container with ID starting with 81a296cd4ef857dfba5b164537ad5f96e8499d2000ff15f91787b1fd4a8e4d7b not found: ID does not exist" Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.176554 4771 scope.go:117] "RemoveContainer" containerID="fd74ef67ecdd38afd00a3e7259fd44fef333a68ad3bf16951d37731e4e68bd26" Oct 02 10:59:12 crc kubenswrapper[4771]: E1002 10:59:12.177038 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd74ef67ecdd38afd00a3e7259fd44fef333a68ad3bf16951d37731e4e68bd26\": container with ID starting with fd74ef67ecdd38afd00a3e7259fd44fef333a68ad3bf16951d37731e4e68bd26 not found: ID does not exist" containerID="fd74ef67ecdd38afd00a3e7259fd44fef333a68ad3bf16951d37731e4e68bd26" Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.177061 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd74ef67ecdd38afd00a3e7259fd44fef333a68ad3bf16951d37731e4e68bd26"} err="failed to get container status \"fd74ef67ecdd38afd00a3e7259fd44fef333a68ad3bf16951d37731e4e68bd26\": rpc error: code = NotFound desc = could not find container \"fd74ef67ecdd38afd00a3e7259fd44fef333a68ad3bf16951d37731e4e68bd26\": container with ID starting with fd74ef67ecdd38afd00a3e7259fd44fef333a68ad3bf16951d37731e4e68bd26 not found: ID does not exist" Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.177077 4771 scope.go:117] "RemoveContainer" containerID="55bcd7d2fac02922cb300b393a92ae8a70e839a9d1278c6e32288d4afabcdf81" Oct 02 10:59:12 crc kubenswrapper[4771]: E1002 10:59:12.177359 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55bcd7d2fac02922cb300b393a92ae8a70e839a9d1278c6e32288d4afabcdf81\": container with ID starting with 55bcd7d2fac02922cb300b393a92ae8a70e839a9d1278c6e32288d4afabcdf81 not found: ID does not exist" containerID="55bcd7d2fac02922cb300b393a92ae8a70e839a9d1278c6e32288d4afabcdf81" Oct 02 10:59:12 crc kubenswrapper[4771]: I1002 10:59:12.177386 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55bcd7d2fac02922cb300b393a92ae8a70e839a9d1278c6e32288d4afabcdf81"} err="failed to get container status \"55bcd7d2fac02922cb300b393a92ae8a70e839a9d1278c6e32288d4afabcdf81\": rpc error: code = NotFound desc = could not find container \"55bcd7d2fac02922cb300b393a92ae8a70e839a9d1278c6e32288d4afabcdf81\": container with ID starting with 55bcd7d2fac02922cb300b393a92ae8a70e839a9d1278c6e32288d4afabcdf81 not found: ID does not exist" Oct 02 10:59:13 crc kubenswrapper[4771]: I1002 10:59:13.700230 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4099631-fe1c-4047-bd5c-c46ab0b808c3" path="/var/lib/kubelet/pods/a4099631-fe1c-4047-bd5c-c46ab0b808c3/volumes" Oct 02 10:59:42 crc kubenswrapper[4771]: I1002 10:59:42.146267 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:59:42 crc kubenswrapper[4771]: I1002 10:59:42.146830 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.098888 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sh2f6"] Oct 02 10:59:48 crc kubenswrapper[4771]: E1002 10:59:48.100029 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4099631-fe1c-4047-bd5c-c46ab0b808c3" containerName="extract-utilities" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.100063 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4099631-fe1c-4047-bd5c-c46ab0b808c3" containerName="extract-utilities" Oct 02 10:59:48 crc kubenswrapper[4771]: E1002 10:59:48.100091 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4099631-fe1c-4047-bd5c-c46ab0b808c3" containerName="registry-server" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.100100 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4099631-fe1c-4047-bd5c-c46ab0b808c3" containerName="registry-server" Oct 02 10:59:48 crc kubenswrapper[4771]: E1002 10:59:48.100156 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4099631-fe1c-4047-bd5c-c46ab0b808c3" containerName="extract-content" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.100166 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4099631-fe1c-4047-bd5c-c46ab0b808c3" containerName="extract-content" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.100517 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4099631-fe1c-4047-bd5c-c46ab0b808c3" containerName="registry-server" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.102727 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.112196 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sh2f6"] Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.253330 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4wtw\" (UniqueName: \"kubernetes.io/projected/27efdb1f-dddb-40cb-a974-f1828b839709-kube-api-access-q4wtw\") pod \"certified-operators-sh2f6\" (UID: \"27efdb1f-dddb-40cb-a974-f1828b839709\") " pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.253677 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27efdb1f-dddb-40cb-a974-f1828b839709-catalog-content\") pod \"certified-operators-sh2f6\" (UID: \"27efdb1f-dddb-40cb-a974-f1828b839709\") " pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.253780 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27efdb1f-dddb-40cb-a974-f1828b839709-utilities\") pod \"certified-operators-sh2f6\" (UID: \"27efdb1f-dddb-40cb-a974-f1828b839709\") " pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.356264 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4wtw\" (UniqueName: \"kubernetes.io/projected/27efdb1f-dddb-40cb-a974-f1828b839709-kube-api-access-q4wtw\") pod \"certified-operators-sh2f6\" (UID: \"27efdb1f-dddb-40cb-a974-f1828b839709\") " pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.356324 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27efdb1f-dddb-40cb-a974-f1828b839709-catalog-content\") pod \"certified-operators-sh2f6\" (UID: \"27efdb1f-dddb-40cb-a974-f1828b839709\") " pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.356492 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27efdb1f-dddb-40cb-a974-f1828b839709-utilities\") pod \"certified-operators-sh2f6\" (UID: \"27efdb1f-dddb-40cb-a974-f1828b839709\") " pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.357018 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27efdb1f-dddb-40cb-a974-f1828b839709-catalog-content\") pod \"certified-operators-sh2f6\" (UID: \"27efdb1f-dddb-40cb-a974-f1828b839709\") " pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.357207 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27efdb1f-dddb-40cb-a974-f1828b839709-utilities\") pod \"certified-operators-sh2f6\" (UID: \"27efdb1f-dddb-40cb-a974-f1828b839709\") " pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.381048 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4wtw\" (UniqueName: \"kubernetes.io/projected/27efdb1f-dddb-40cb-a974-f1828b839709-kube-api-access-q4wtw\") pod \"certified-operators-sh2f6\" (UID: \"27efdb1f-dddb-40cb-a974-f1828b839709\") " pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:48 crc kubenswrapper[4771]: I1002 10:59:48.432639 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:49 crc kubenswrapper[4771]: I1002 10:59:49.066661 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sh2f6"] Oct 02 10:59:49 crc kubenswrapper[4771]: I1002 10:59:49.482114 4771 generic.go:334] "Generic (PLEG): container finished" podID="27efdb1f-dddb-40cb-a974-f1828b839709" containerID="043ce663a8055d0a2b9f612833a514dbafc20bca35053ced6d8f161bcb307abd" exitCode=0 Oct 02 10:59:49 crc kubenswrapper[4771]: I1002 10:59:49.482413 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh2f6" event={"ID":"27efdb1f-dddb-40cb-a974-f1828b839709","Type":"ContainerDied","Data":"043ce663a8055d0a2b9f612833a514dbafc20bca35053ced6d8f161bcb307abd"} Oct 02 10:59:49 crc kubenswrapper[4771]: I1002 10:59:49.482438 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh2f6" event={"ID":"27efdb1f-dddb-40cb-a974-f1828b839709","Type":"ContainerStarted","Data":"8c153332f951ea3bf9b1b051967a20b0d6d08b06a8fd3a9e7fe65bff57aeef89"} Oct 02 10:59:51 crc kubenswrapper[4771]: I1002 10:59:51.504227 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh2f6" event={"ID":"27efdb1f-dddb-40cb-a974-f1828b839709","Type":"ContainerStarted","Data":"a1be5472f1a1e474534aa86daff977abbc811675a76fed2bec6e5cf3fb4ea782"} Oct 02 10:59:52 crc kubenswrapper[4771]: I1002 10:59:52.516268 4771 generic.go:334] "Generic (PLEG): container finished" podID="27efdb1f-dddb-40cb-a974-f1828b839709" containerID="a1be5472f1a1e474534aa86daff977abbc811675a76fed2bec6e5cf3fb4ea782" exitCode=0 Oct 02 10:59:52 crc kubenswrapper[4771]: I1002 10:59:52.516317 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh2f6" event={"ID":"27efdb1f-dddb-40cb-a974-f1828b839709","Type":"ContainerDied","Data":"a1be5472f1a1e474534aa86daff977abbc811675a76fed2bec6e5cf3fb4ea782"} Oct 02 10:59:54 crc kubenswrapper[4771]: I1002 10:59:54.540869 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh2f6" event={"ID":"27efdb1f-dddb-40cb-a974-f1828b839709","Type":"ContainerStarted","Data":"63c84d3f0fa4a09033237368caae0d045c64b5cd5e2df3b19844040fc55f8515"} Oct 02 10:59:54 crc kubenswrapper[4771]: I1002 10:59:54.560546 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sh2f6" podStartSLOduration=3.002489549 podStartE2EDuration="6.560528096s" podCreationTimestamp="2025-10-02 10:59:48 +0000 UTC" firstStartedPulling="2025-10-02 10:59:49.484216254 +0000 UTC m=+4977.131901321" lastFinishedPulling="2025-10-02 10:59:53.042254801 +0000 UTC m=+4980.689939868" observedRunningTime="2025-10-02 10:59:54.55805656 +0000 UTC m=+4982.205741627" watchObservedRunningTime="2025-10-02 10:59:54.560528096 +0000 UTC m=+4982.208213163" Oct 02 10:59:58 crc kubenswrapper[4771]: I1002 10:59:58.433224 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:58 crc kubenswrapper[4771]: I1002 10:59:58.433758 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:58 crc kubenswrapper[4771]: I1002 10:59:58.490630 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:58 crc kubenswrapper[4771]: I1002 10:59:58.632501 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 10:59:58 crc kubenswrapper[4771]: I1002 10:59:58.726540 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sh2f6"] Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.149005 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9"] Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.151520 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.155061 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.155328 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.172797 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9"] Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.253242 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zprmw\" (UniqueName: \"kubernetes.io/projected/a89c2bce-3833-4930-9ac9-b82352e42f0f-kube-api-access-zprmw\") pod \"collect-profiles-29323380-8zld9\" (UID: \"a89c2bce-3833-4930-9ac9-b82352e42f0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.253515 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a89c2bce-3833-4930-9ac9-b82352e42f0f-config-volume\") pod \"collect-profiles-29323380-8zld9\" (UID: \"a89c2bce-3833-4930-9ac9-b82352e42f0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.253598 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a89c2bce-3833-4930-9ac9-b82352e42f0f-secret-volume\") pod \"collect-profiles-29323380-8zld9\" (UID: \"a89c2bce-3833-4930-9ac9-b82352e42f0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.355542 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zprmw\" (UniqueName: \"kubernetes.io/projected/a89c2bce-3833-4930-9ac9-b82352e42f0f-kube-api-access-zprmw\") pod \"collect-profiles-29323380-8zld9\" (UID: \"a89c2bce-3833-4930-9ac9-b82352e42f0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.355700 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a89c2bce-3833-4930-9ac9-b82352e42f0f-config-volume\") pod \"collect-profiles-29323380-8zld9\" (UID: \"a89c2bce-3833-4930-9ac9-b82352e42f0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.355752 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a89c2bce-3833-4930-9ac9-b82352e42f0f-secret-volume\") pod \"collect-profiles-29323380-8zld9\" (UID: \"a89c2bce-3833-4930-9ac9-b82352e42f0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.356775 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a89c2bce-3833-4930-9ac9-b82352e42f0f-config-volume\") pod \"collect-profiles-29323380-8zld9\" (UID: \"a89c2bce-3833-4930-9ac9-b82352e42f0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.366053 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a89c2bce-3833-4930-9ac9-b82352e42f0f-secret-volume\") pod \"collect-profiles-29323380-8zld9\" (UID: \"a89c2bce-3833-4930-9ac9-b82352e42f0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.375335 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zprmw\" (UniqueName: \"kubernetes.io/projected/a89c2bce-3833-4930-9ac9-b82352e42f0f-kube-api-access-zprmw\") pod \"collect-profiles-29323380-8zld9\" (UID: \"a89c2bce-3833-4930-9ac9-b82352e42f0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.474667 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" Oct 02 11:00:00 crc kubenswrapper[4771]: I1002 11:00:00.607064 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sh2f6" podUID="27efdb1f-dddb-40cb-a974-f1828b839709" containerName="registry-server" containerID="cri-o://63c84d3f0fa4a09033237368caae0d045c64b5cd5e2df3b19844040fc55f8515" gracePeriod=2 Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.060414 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9"] Oct 02 11:00:01 crc kubenswrapper[4771]: W1002 11:00:01.087534 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda89c2bce_3833_4930_9ac9_b82352e42f0f.slice/crio-c17300bba4017a10452cb71c0da707a7adc83e6adf81088ccde27f6ae1ac2db7 WatchSource:0}: Error finding container c17300bba4017a10452cb71c0da707a7adc83e6adf81088ccde27f6ae1ac2db7: Status 404 returned error can't find the container with id c17300bba4017a10452cb71c0da707a7adc83e6adf81088ccde27f6ae1ac2db7 Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.104257 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.176564 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4wtw\" (UniqueName: \"kubernetes.io/projected/27efdb1f-dddb-40cb-a974-f1828b839709-kube-api-access-q4wtw\") pod \"27efdb1f-dddb-40cb-a974-f1828b839709\" (UID: \"27efdb1f-dddb-40cb-a974-f1828b839709\") " Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.176666 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27efdb1f-dddb-40cb-a974-f1828b839709-utilities\") pod \"27efdb1f-dddb-40cb-a974-f1828b839709\" (UID: \"27efdb1f-dddb-40cb-a974-f1828b839709\") " Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.176927 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27efdb1f-dddb-40cb-a974-f1828b839709-catalog-content\") pod \"27efdb1f-dddb-40cb-a974-f1828b839709\" (UID: \"27efdb1f-dddb-40cb-a974-f1828b839709\") " Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.186753 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27efdb1f-dddb-40cb-a974-f1828b839709-utilities" (OuterVolumeSpecName: "utilities") pod "27efdb1f-dddb-40cb-a974-f1828b839709" (UID: "27efdb1f-dddb-40cb-a974-f1828b839709"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.193121 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27efdb1f-dddb-40cb-a974-f1828b839709-kube-api-access-q4wtw" (OuterVolumeSpecName: "kube-api-access-q4wtw") pod "27efdb1f-dddb-40cb-a974-f1828b839709" (UID: "27efdb1f-dddb-40cb-a974-f1828b839709"). InnerVolumeSpecName "kube-api-access-q4wtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.272415 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27efdb1f-dddb-40cb-a974-f1828b839709-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27efdb1f-dddb-40cb-a974-f1828b839709" (UID: "27efdb1f-dddb-40cb-a974-f1828b839709"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.280747 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27efdb1f-dddb-40cb-a974-f1828b839709-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.280788 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4wtw\" (UniqueName: \"kubernetes.io/projected/27efdb1f-dddb-40cb-a974-f1828b839709-kube-api-access-q4wtw\") on node \"crc\" DevicePath \"\"" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.280831 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27efdb1f-dddb-40cb-a974-f1828b839709-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.617723 4771 generic.go:334] "Generic (PLEG): container finished" podID="a89c2bce-3833-4930-9ac9-b82352e42f0f" containerID="c8bf9cf96c4f0287badd4798f71f35912aa0b9463fcfe5ce7f49b1f3f569a533" exitCode=0 Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.617771 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" event={"ID":"a89c2bce-3833-4930-9ac9-b82352e42f0f","Type":"ContainerDied","Data":"c8bf9cf96c4f0287badd4798f71f35912aa0b9463fcfe5ce7f49b1f3f569a533"} Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.618087 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" event={"ID":"a89c2bce-3833-4930-9ac9-b82352e42f0f","Type":"ContainerStarted","Data":"c17300bba4017a10452cb71c0da707a7adc83e6adf81088ccde27f6ae1ac2db7"} Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.621780 4771 generic.go:334] "Generic (PLEG): container finished" podID="27efdb1f-dddb-40cb-a974-f1828b839709" containerID="63c84d3f0fa4a09033237368caae0d045c64b5cd5e2df3b19844040fc55f8515" exitCode=0 Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.621827 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh2f6" event={"ID":"27efdb1f-dddb-40cb-a974-f1828b839709","Type":"ContainerDied","Data":"63c84d3f0fa4a09033237368caae0d045c64b5cd5e2df3b19844040fc55f8515"} Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.621854 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sh2f6" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.621875 4771 scope.go:117] "RemoveContainer" containerID="63c84d3f0fa4a09033237368caae0d045c64b5cd5e2df3b19844040fc55f8515" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.621861 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sh2f6" event={"ID":"27efdb1f-dddb-40cb-a974-f1828b839709","Type":"ContainerDied","Data":"8c153332f951ea3bf9b1b051967a20b0d6d08b06a8fd3a9e7fe65bff57aeef89"} Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.655966 4771 scope.go:117] "RemoveContainer" containerID="a1be5472f1a1e474534aa86daff977abbc811675a76fed2bec6e5cf3fb4ea782" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.677731 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sh2f6"] Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.683236 4771 scope.go:117] "RemoveContainer" containerID="043ce663a8055d0a2b9f612833a514dbafc20bca35053ced6d8f161bcb307abd" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.703610 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sh2f6"] Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.756998 4771 scope.go:117] "RemoveContainer" containerID="63c84d3f0fa4a09033237368caae0d045c64b5cd5e2df3b19844040fc55f8515" Oct 02 11:00:01 crc kubenswrapper[4771]: E1002 11:00:01.757468 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63c84d3f0fa4a09033237368caae0d045c64b5cd5e2df3b19844040fc55f8515\": container with ID starting with 63c84d3f0fa4a09033237368caae0d045c64b5cd5e2df3b19844040fc55f8515 not found: ID does not exist" containerID="63c84d3f0fa4a09033237368caae0d045c64b5cd5e2df3b19844040fc55f8515" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.757499 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63c84d3f0fa4a09033237368caae0d045c64b5cd5e2df3b19844040fc55f8515"} err="failed to get container status \"63c84d3f0fa4a09033237368caae0d045c64b5cd5e2df3b19844040fc55f8515\": rpc error: code = NotFound desc = could not find container \"63c84d3f0fa4a09033237368caae0d045c64b5cd5e2df3b19844040fc55f8515\": container with ID starting with 63c84d3f0fa4a09033237368caae0d045c64b5cd5e2df3b19844040fc55f8515 not found: ID does not exist" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.757520 4771 scope.go:117] "RemoveContainer" containerID="a1be5472f1a1e474534aa86daff977abbc811675a76fed2bec6e5cf3fb4ea782" Oct 02 11:00:01 crc kubenswrapper[4771]: E1002 11:00:01.757904 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1be5472f1a1e474534aa86daff977abbc811675a76fed2bec6e5cf3fb4ea782\": container with ID starting with a1be5472f1a1e474534aa86daff977abbc811675a76fed2bec6e5cf3fb4ea782 not found: ID does not exist" containerID="a1be5472f1a1e474534aa86daff977abbc811675a76fed2bec6e5cf3fb4ea782" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.757924 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1be5472f1a1e474534aa86daff977abbc811675a76fed2bec6e5cf3fb4ea782"} err="failed to get container status \"a1be5472f1a1e474534aa86daff977abbc811675a76fed2bec6e5cf3fb4ea782\": rpc error: code = NotFound desc = could not find container \"a1be5472f1a1e474534aa86daff977abbc811675a76fed2bec6e5cf3fb4ea782\": container with ID starting with a1be5472f1a1e474534aa86daff977abbc811675a76fed2bec6e5cf3fb4ea782 not found: ID does not exist" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.757936 4771 scope.go:117] "RemoveContainer" containerID="043ce663a8055d0a2b9f612833a514dbafc20bca35053ced6d8f161bcb307abd" Oct 02 11:00:01 crc kubenswrapper[4771]: E1002 11:00:01.758350 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"043ce663a8055d0a2b9f612833a514dbafc20bca35053ced6d8f161bcb307abd\": container with ID starting with 043ce663a8055d0a2b9f612833a514dbafc20bca35053ced6d8f161bcb307abd not found: ID does not exist" containerID="043ce663a8055d0a2b9f612833a514dbafc20bca35053ced6d8f161bcb307abd" Oct 02 11:00:01 crc kubenswrapper[4771]: I1002 11:00:01.758398 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"043ce663a8055d0a2b9f612833a514dbafc20bca35053ced6d8f161bcb307abd"} err="failed to get container status \"043ce663a8055d0a2b9f612833a514dbafc20bca35053ced6d8f161bcb307abd\": rpc error: code = NotFound desc = could not find container \"043ce663a8055d0a2b9f612833a514dbafc20bca35053ced6d8f161bcb307abd\": container with ID starting with 043ce663a8055d0a2b9f612833a514dbafc20bca35053ced6d8f161bcb307abd not found: ID does not exist" Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.052742 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.123120 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a89c2bce-3833-4930-9ac9-b82352e42f0f-config-volume\") pod \"a89c2bce-3833-4930-9ac9-b82352e42f0f\" (UID: \"a89c2bce-3833-4930-9ac9-b82352e42f0f\") " Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.123232 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a89c2bce-3833-4930-9ac9-b82352e42f0f-secret-volume\") pod \"a89c2bce-3833-4930-9ac9-b82352e42f0f\" (UID: \"a89c2bce-3833-4930-9ac9-b82352e42f0f\") " Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.123323 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zprmw\" (UniqueName: \"kubernetes.io/projected/a89c2bce-3833-4930-9ac9-b82352e42f0f-kube-api-access-zprmw\") pod \"a89c2bce-3833-4930-9ac9-b82352e42f0f\" (UID: \"a89c2bce-3833-4930-9ac9-b82352e42f0f\") " Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.125932 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a89c2bce-3833-4930-9ac9-b82352e42f0f-config-volume" (OuterVolumeSpecName: "config-volume") pod "a89c2bce-3833-4930-9ac9-b82352e42f0f" (UID: "a89c2bce-3833-4930-9ac9-b82352e42f0f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.135340 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a89c2bce-3833-4930-9ac9-b82352e42f0f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a89c2bce-3833-4930-9ac9-b82352e42f0f" (UID: "a89c2bce-3833-4930-9ac9-b82352e42f0f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.135484 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a89c2bce-3833-4930-9ac9-b82352e42f0f-kube-api-access-zprmw" (OuterVolumeSpecName: "kube-api-access-zprmw") pod "a89c2bce-3833-4930-9ac9-b82352e42f0f" (UID: "a89c2bce-3833-4930-9ac9-b82352e42f0f"). InnerVolumeSpecName "kube-api-access-zprmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.225915 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zprmw\" (UniqueName: \"kubernetes.io/projected/a89c2bce-3833-4930-9ac9-b82352e42f0f-kube-api-access-zprmw\") on node \"crc\" DevicePath \"\"" Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.225957 4771 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a89c2bce-3833-4930-9ac9-b82352e42f0f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.225984 4771 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a89c2bce-3833-4930-9ac9-b82352e42f0f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.651339 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" event={"ID":"a89c2bce-3833-4930-9ac9-b82352e42f0f","Type":"ContainerDied","Data":"c17300bba4017a10452cb71c0da707a7adc83e6adf81088ccde27f6ae1ac2db7"} Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.651695 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c17300bba4017a10452cb71c0da707a7adc83e6adf81088ccde27f6ae1ac2db7" Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.651833 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9" Oct 02 11:00:03 crc kubenswrapper[4771]: I1002 11:00:03.706399 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27efdb1f-dddb-40cb-a974-f1828b839709" path="/var/lib/kubelet/pods/27efdb1f-dddb-40cb-a974-f1828b839709/volumes" Oct 02 11:00:04 crc kubenswrapper[4771]: I1002 11:00:04.136760 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9"] Oct 02 11:00:04 crc kubenswrapper[4771]: I1002 11:00:04.147384 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323335-wt5h9"] Oct 02 11:00:05 crc kubenswrapper[4771]: I1002 11:00:05.700011 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9fc2c00-9a21-4c8e-ac5f-de3e660aa217" path="/var/lib/kubelet/pods/a9fc2c00-9a21-4c8e-ac5f-de3e660aa217/volumes" Oct 02 11:00:05 crc kubenswrapper[4771]: I1002 11:00:05.957235 4771 scope.go:117] "RemoveContainer" containerID="62529b0c27dd7e5451357bd026448129da580c3778f5fe72162f8c158870df52" Oct 02 11:00:12 crc kubenswrapper[4771]: I1002 11:00:12.146011 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:00:12 crc kubenswrapper[4771]: I1002 11:00:12.146521 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:00:42 crc kubenswrapper[4771]: I1002 11:00:42.146022 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:00:42 crc kubenswrapper[4771]: I1002 11:00:42.147109 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:00:42 crc kubenswrapper[4771]: I1002 11:00:42.147211 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 11:00:42 crc kubenswrapper[4771]: I1002 11:00:42.148902 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:00:42 crc kubenswrapper[4771]: I1002 11:00:42.148983 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" gracePeriod=600 Oct 02 11:00:42 crc kubenswrapper[4771]: E1002 11:00:42.286997 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:00:43 crc kubenswrapper[4771]: I1002 11:00:43.113389 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" exitCode=0 Oct 02 11:00:43 crc kubenswrapper[4771]: I1002 11:00:43.113416 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b"} Oct 02 11:00:43 crc kubenswrapper[4771]: I1002 11:00:43.114035 4771 scope.go:117] "RemoveContainer" containerID="3879c77192f81ec753429c0f3f323cb067948912c2feb70ad2e6e8689e44370a" Oct 02 11:00:43 crc kubenswrapper[4771]: I1002 11:00:43.114870 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:00:43 crc kubenswrapper[4771]: E1002 11:00:43.115263 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:00:53 crc kubenswrapper[4771]: I1002 11:00:53.698571 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:00:53 crc kubenswrapper[4771]: E1002 11:00:53.699609 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.153537 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29323381-2gw54"] Oct 02 11:01:00 crc kubenswrapper[4771]: E1002 11:01:00.154619 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27efdb1f-dddb-40cb-a974-f1828b839709" containerName="registry-server" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.154633 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="27efdb1f-dddb-40cb-a974-f1828b839709" containerName="registry-server" Oct 02 11:01:00 crc kubenswrapper[4771]: E1002 11:01:00.154664 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27efdb1f-dddb-40cb-a974-f1828b839709" containerName="extract-content" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.154670 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="27efdb1f-dddb-40cb-a974-f1828b839709" containerName="extract-content" Oct 02 11:01:00 crc kubenswrapper[4771]: E1002 11:01:00.154680 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a89c2bce-3833-4930-9ac9-b82352e42f0f" containerName="collect-profiles" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.154687 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a89c2bce-3833-4930-9ac9-b82352e42f0f" containerName="collect-profiles" Oct 02 11:01:00 crc kubenswrapper[4771]: E1002 11:01:00.154710 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27efdb1f-dddb-40cb-a974-f1828b839709" containerName="extract-utilities" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.154715 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="27efdb1f-dddb-40cb-a974-f1828b839709" containerName="extract-utilities" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.154977 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a89c2bce-3833-4930-9ac9-b82352e42f0f" containerName="collect-profiles" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.154996 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="27efdb1f-dddb-40cb-a974-f1828b839709" containerName="registry-server" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.155877 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.165995 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323381-2gw54"] Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.238687 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dnpl\" (UniqueName: \"kubernetes.io/projected/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-kube-api-access-6dnpl\") pod \"keystone-cron-29323381-2gw54\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.238797 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-config-data\") pod \"keystone-cron-29323381-2gw54\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.238825 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-fernet-keys\") pod \"keystone-cron-29323381-2gw54\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.238981 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-combined-ca-bundle\") pod \"keystone-cron-29323381-2gw54\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.340684 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-fernet-keys\") pod \"keystone-cron-29323381-2gw54\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.340859 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-combined-ca-bundle\") pod \"keystone-cron-29323381-2gw54\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.341071 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dnpl\" (UniqueName: \"kubernetes.io/projected/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-kube-api-access-6dnpl\") pod \"keystone-cron-29323381-2gw54\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.341219 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-config-data\") pod \"keystone-cron-29323381-2gw54\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.348979 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-combined-ca-bundle\") pod \"keystone-cron-29323381-2gw54\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.351051 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-fernet-keys\") pod \"keystone-cron-29323381-2gw54\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.352203 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-config-data\") pod \"keystone-cron-29323381-2gw54\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.367454 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dnpl\" (UniqueName: \"kubernetes.io/projected/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-kube-api-access-6dnpl\") pod \"keystone-cron-29323381-2gw54\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:00 crc kubenswrapper[4771]: I1002 11:01:00.498479 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:01 crc kubenswrapper[4771]: I1002 11:01:01.033311 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323381-2gw54"] Oct 02 11:01:01 crc kubenswrapper[4771]: I1002 11:01:01.382518 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323381-2gw54" event={"ID":"6c79e9b7-b296-47b5-a27b-e72cbf77adf8","Type":"ContainerStarted","Data":"7af62075dccc4ed64bb76c02539cc4a1e04485ae0810e21f66753c047ca32741"} Oct 02 11:01:01 crc kubenswrapper[4771]: I1002 11:01:01.382559 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323381-2gw54" event={"ID":"6c79e9b7-b296-47b5-a27b-e72cbf77adf8","Type":"ContainerStarted","Data":"94fc478768964720fcbc9d4a94aa05331b9ac79d9625c6c08e087ebf81284bd4"} Oct 02 11:01:01 crc kubenswrapper[4771]: I1002 11:01:01.399314 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29323381-2gw54" podStartSLOduration=1.399283086 podStartE2EDuration="1.399283086s" podCreationTimestamp="2025-10-02 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:01:01.39792243 +0000 UTC m=+5049.045607497" watchObservedRunningTime="2025-10-02 11:01:01.399283086 +0000 UTC m=+5049.046968153" Oct 02 11:01:05 crc kubenswrapper[4771]: I1002 11:01:05.425489 4771 generic.go:334] "Generic (PLEG): container finished" podID="6c79e9b7-b296-47b5-a27b-e72cbf77adf8" containerID="7af62075dccc4ed64bb76c02539cc4a1e04485ae0810e21f66753c047ca32741" exitCode=0 Oct 02 11:01:05 crc kubenswrapper[4771]: I1002 11:01:05.425570 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323381-2gw54" event={"ID":"6c79e9b7-b296-47b5-a27b-e72cbf77adf8","Type":"ContainerDied","Data":"7af62075dccc4ed64bb76c02539cc4a1e04485ae0810e21f66753c047ca32741"} Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.447254 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323381-2gw54" event={"ID":"6c79e9b7-b296-47b5-a27b-e72cbf77adf8","Type":"ContainerDied","Data":"94fc478768964720fcbc9d4a94aa05331b9ac79d9625c6c08e087ebf81284bd4"} Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.447671 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94fc478768964720fcbc9d4a94aa05331b9ac79d9625c6c08e087ebf81284bd4" Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.504996 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.620659 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-combined-ca-bundle\") pod \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.620817 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-config-data\") pod \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.620959 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dnpl\" (UniqueName: \"kubernetes.io/projected/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-kube-api-access-6dnpl\") pod \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.621054 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-fernet-keys\") pod \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\" (UID: \"6c79e9b7-b296-47b5-a27b-e72cbf77adf8\") " Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.626584 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6c79e9b7-b296-47b5-a27b-e72cbf77adf8" (UID: "6c79e9b7-b296-47b5-a27b-e72cbf77adf8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.626804 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-kube-api-access-6dnpl" (OuterVolumeSpecName: "kube-api-access-6dnpl") pod "6c79e9b7-b296-47b5-a27b-e72cbf77adf8" (UID: "6c79e9b7-b296-47b5-a27b-e72cbf77adf8"). InnerVolumeSpecName "kube-api-access-6dnpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.672449 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c79e9b7-b296-47b5-a27b-e72cbf77adf8" (UID: "6c79e9b7-b296-47b5-a27b-e72cbf77adf8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.725019 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.725051 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dnpl\" (UniqueName: \"kubernetes.io/projected/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-kube-api-access-6dnpl\") on node \"crc\" DevicePath \"\"" Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.725064 4771 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.796631 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-config-data" (OuterVolumeSpecName: "config-data") pod "6c79e9b7-b296-47b5-a27b-e72cbf77adf8" (UID: "6c79e9b7-b296-47b5-a27b-e72cbf77adf8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:01:07 crc kubenswrapper[4771]: I1002 11:01:07.827872 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c79e9b7-b296-47b5-a27b-e72cbf77adf8-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:01:08 crc kubenswrapper[4771]: I1002 11:01:08.459606 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323381-2gw54" Oct 02 11:01:08 crc kubenswrapper[4771]: I1002 11:01:08.683937 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:01:08 crc kubenswrapper[4771]: E1002 11:01:08.684414 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:01:21 crc kubenswrapper[4771]: I1002 11:01:21.682161 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:01:21 crc kubenswrapper[4771]: E1002 11:01:21.683041 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:01:36 crc kubenswrapper[4771]: I1002 11:01:36.681708 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:01:36 crc kubenswrapper[4771]: E1002 11:01:36.682844 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:01:49 crc kubenswrapper[4771]: I1002 11:01:49.681345 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:01:49 crc kubenswrapper[4771]: E1002 11:01:49.682189 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:02:03 crc kubenswrapper[4771]: I1002 11:02:03.690357 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:02:03 crc kubenswrapper[4771]: E1002 11:02:03.691500 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:02:18 crc kubenswrapper[4771]: I1002 11:02:18.681787 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:02:18 crc kubenswrapper[4771]: E1002 11:02:18.683090 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:02:30 crc kubenswrapper[4771]: I1002 11:02:30.681379 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:02:30 crc kubenswrapper[4771]: E1002 11:02:30.682296 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:02:43 crc kubenswrapper[4771]: I1002 11:02:43.692584 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:02:43 crc kubenswrapper[4771]: E1002 11:02:43.693587 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:02:55 crc kubenswrapper[4771]: I1002 11:02:55.681758 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:02:55 crc kubenswrapper[4771]: E1002 11:02:55.682780 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:03:07 crc kubenswrapper[4771]: I1002 11:03:07.682560 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:03:07 crc kubenswrapper[4771]: E1002 11:03:07.683664 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:03:19 crc kubenswrapper[4771]: I1002 11:03:19.681358 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:03:19 crc kubenswrapper[4771]: E1002 11:03:19.682232 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:03:30 crc kubenswrapper[4771]: I1002 11:03:30.681225 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:03:30 crc kubenswrapper[4771]: E1002 11:03:30.682144 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:03:45 crc kubenswrapper[4771]: I1002 11:03:45.681740 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:03:45 crc kubenswrapper[4771]: E1002 11:03:45.682724 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:04:00 crc kubenswrapper[4771]: I1002 11:04:00.682170 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:04:00 crc kubenswrapper[4771]: E1002 11:04:00.683141 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:04:12 crc kubenswrapper[4771]: I1002 11:04:12.682100 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:04:12 crc kubenswrapper[4771]: E1002 11:04:12.682964 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:04:25 crc kubenswrapper[4771]: I1002 11:04:25.683075 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:04:25 crc kubenswrapper[4771]: E1002 11:04:25.683958 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:04:36 crc kubenswrapper[4771]: I1002 11:04:36.682233 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:04:36 crc kubenswrapper[4771]: E1002 11:04:36.683570 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:04:48 crc kubenswrapper[4771]: I1002 11:04:48.682810 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:04:48 crc kubenswrapper[4771]: E1002 11:04:48.683819 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:04:53 crc kubenswrapper[4771]: I1002 11:04:53.851003 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7f847"] Oct 02 11:04:53 crc kubenswrapper[4771]: E1002 11:04:53.852814 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c79e9b7-b296-47b5-a27b-e72cbf77adf8" containerName="keystone-cron" Oct 02 11:04:53 crc kubenswrapper[4771]: I1002 11:04:53.852832 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c79e9b7-b296-47b5-a27b-e72cbf77adf8" containerName="keystone-cron" Oct 02 11:04:53 crc kubenswrapper[4771]: I1002 11:04:53.853094 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c79e9b7-b296-47b5-a27b-e72cbf77adf8" containerName="keystone-cron" Oct 02 11:04:53 crc kubenswrapper[4771]: I1002 11:04:53.857277 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:04:53 crc kubenswrapper[4771]: I1002 11:04:53.868262 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7f847"] Oct 02 11:04:54 crc kubenswrapper[4771]: I1002 11:04:54.025484 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1650a38-f0a8-43ad-b7db-babf835c0199-catalog-content\") pod \"redhat-operators-7f847\" (UID: \"b1650a38-f0a8-43ad-b7db-babf835c0199\") " pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:04:54 crc kubenswrapper[4771]: I1002 11:04:54.026038 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp6lz\" (UniqueName: \"kubernetes.io/projected/b1650a38-f0a8-43ad-b7db-babf835c0199-kube-api-access-jp6lz\") pod \"redhat-operators-7f847\" (UID: \"b1650a38-f0a8-43ad-b7db-babf835c0199\") " pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:04:54 crc kubenswrapper[4771]: I1002 11:04:54.026266 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1650a38-f0a8-43ad-b7db-babf835c0199-utilities\") pod \"redhat-operators-7f847\" (UID: \"b1650a38-f0a8-43ad-b7db-babf835c0199\") " pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:04:54 crc kubenswrapper[4771]: I1002 11:04:54.128567 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1650a38-f0a8-43ad-b7db-babf835c0199-catalog-content\") pod \"redhat-operators-7f847\" (UID: \"b1650a38-f0a8-43ad-b7db-babf835c0199\") " pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:04:54 crc kubenswrapper[4771]: I1002 11:04:54.128643 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp6lz\" (UniqueName: \"kubernetes.io/projected/b1650a38-f0a8-43ad-b7db-babf835c0199-kube-api-access-jp6lz\") pod \"redhat-operators-7f847\" (UID: \"b1650a38-f0a8-43ad-b7db-babf835c0199\") " pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:04:54 crc kubenswrapper[4771]: I1002 11:04:54.128769 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1650a38-f0a8-43ad-b7db-babf835c0199-utilities\") pod \"redhat-operators-7f847\" (UID: \"b1650a38-f0a8-43ad-b7db-babf835c0199\") " pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:04:54 crc kubenswrapper[4771]: I1002 11:04:54.129068 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1650a38-f0a8-43ad-b7db-babf835c0199-catalog-content\") pod \"redhat-operators-7f847\" (UID: \"b1650a38-f0a8-43ad-b7db-babf835c0199\") " pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:04:54 crc kubenswrapper[4771]: I1002 11:04:54.129112 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1650a38-f0a8-43ad-b7db-babf835c0199-utilities\") pod \"redhat-operators-7f847\" (UID: \"b1650a38-f0a8-43ad-b7db-babf835c0199\") " pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:04:54 crc kubenswrapper[4771]: I1002 11:04:54.159604 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp6lz\" (UniqueName: \"kubernetes.io/projected/b1650a38-f0a8-43ad-b7db-babf835c0199-kube-api-access-jp6lz\") pod \"redhat-operators-7f847\" (UID: \"b1650a38-f0a8-43ad-b7db-babf835c0199\") " pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:04:54 crc kubenswrapper[4771]: I1002 11:04:54.200748 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:04:54 crc kubenswrapper[4771]: I1002 11:04:54.762974 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7f847"] Oct 02 11:04:55 crc kubenswrapper[4771]: I1002 11:04:55.208363 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f847" event={"ID":"b1650a38-f0a8-43ad-b7db-babf835c0199","Type":"ContainerStarted","Data":"6ce12b8321d3f5930f53ef475ea3c7567aea8d0b383c33759521968bd8265d56"} Oct 02 11:04:56 crc kubenswrapper[4771]: I1002 11:04:56.227056 4771 generic.go:334] "Generic (PLEG): container finished" podID="b1650a38-f0a8-43ad-b7db-babf835c0199" containerID="2659abaaf62f8054de0e9c814129ec3a7aecf1f9491abcd96afd55eba6215fe8" exitCode=0 Oct 02 11:04:56 crc kubenswrapper[4771]: I1002 11:04:56.227163 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f847" event={"ID":"b1650a38-f0a8-43ad-b7db-babf835c0199","Type":"ContainerDied","Data":"2659abaaf62f8054de0e9c814129ec3a7aecf1f9491abcd96afd55eba6215fe8"} Oct 02 11:04:56 crc kubenswrapper[4771]: I1002 11:04:56.230050 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:04:58 crc kubenswrapper[4771]: I1002 11:04:58.276521 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f847" event={"ID":"b1650a38-f0a8-43ad-b7db-babf835c0199","Type":"ContainerStarted","Data":"b0df8f95cc95d07439ffa5c44dd9905e38989cef7a27a9a705aa63f2835c1dbd"} Oct 02 11:05:01 crc kubenswrapper[4771]: I1002 11:05:01.310990 4771 generic.go:334] "Generic (PLEG): container finished" podID="b1650a38-f0a8-43ad-b7db-babf835c0199" containerID="b0df8f95cc95d07439ffa5c44dd9905e38989cef7a27a9a705aa63f2835c1dbd" exitCode=0 Oct 02 11:05:01 crc kubenswrapper[4771]: I1002 11:05:01.311312 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f847" event={"ID":"b1650a38-f0a8-43ad-b7db-babf835c0199","Type":"ContainerDied","Data":"b0df8f95cc95d07439ffa5c44dd9905e38989cef7a27a9a705aa63f2835c1dbd"} Oct 02 11:05:02 crc kubenswrapper[4771]: I1002 11:05:02.325444 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f847" event={"ID":"b1650a38-f0a8-43ad-b7db-babf835c0199","Type":"ContainerStarted","Data":"dd2a4f049608bf3192fa2913c29ec7d4f18f1a732457ff7b45c3b9a06ddcdc4e"} Oct 02 11:05:02 crc kubenswrapper[4771]: I1002 11:05:02.349028 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7f847" podStartSLOduration=3.70623349 podStartE2EDuration="9.349009421s" podCreationTimestamp="2025-10-02 11:04:53 +0000 UTC" firstStartedPulling="2025-10-02 11:04:56.229678064 +0000 UTC m=+5283.877363151" lastFinishedPulling="2025-10-02 11:05:01.872454015 +0000 UTC m=+5289.520139082" observedRunningTime="2025-10-02 11:05:02.347212204 +0000 UTC m=+5289.994897271" watchObservedRunningTime="2025-10-02 11:05:02.349009421 +0000 UTC m=+5289.996694498" Oct 02 11:05:02 crc kubenswrapper[4771]: I1002 11:05:02.681452 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:05:02 crc kubenswrapper[4771]: E1002 11:05:02.684958 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:05:04 crc kubenswrapper[4771]: I1002 11:05:04.200891 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:05:04 crc kubenswrapper[4771]: I1002 11:05:04.202015 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:05:05 crc kubenswrapper[4771]: I1002 11:05:05.258779 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7f847" podUID="b1650a38-f0a8-43ad-b7db-babf835c0199" containerName="registry-server" probeResult="failure" output=< Oct 02 11:05:05 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:05:05 crc kubenswrapper[4771]: > Oct 02 11:05:14 crc kubenswrapper[4771]: I1002 11:05:14.249279 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:05:14 crc kubenswrapper[4771]: I1002 11:05:14.298789 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:05:14 crc kubenswrapper[4771]: I1002 11:05:14.483565 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7f847"] Oct 02 11:05:14 crc kubenswrapper[4771]: I1002 11:05:14.681068 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:05:14 crc kubenswrapper[4771]: E1002 11:05:14.681434 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:05:15 crc kubenswrapper[4771]: I1002 11:05:15.457002 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7f847" podUID="b1650a38-f0a8-43ad-b7db-babf835c0199" containerName="registry-server" containerID="cri-o://dd2a4f049608bf3192fa2913c29ec7d4f18f1a732457ff7b45c3b9a06ddcdc4e" gracePeriod=2 Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.227911 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.295493 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jp6lz\" (UniqueName: \"kubernetes.io/projected/b1650a38-f0a8-43ad-b7db-babf835c0199-kube-api-access-jp6lz\") pod \"b1650a38-f0a8-43ad-b7db-babf835c0199\" (UID: \"b1650a38-f0a8-43ad-b7db-babf835c0199\") " Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.295572 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1650a38-f0a8-43ad-b7db-babf835c0199-utilities\") pod \"b1650a38-f0a8-43ad-b7db-babf835c0199\" (UID: \"b1650a38-f0a8-43ad-b7db-babf835c0199\") " Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.295851 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1650a38-f0a8-43ad-b7db-babf835c0199-catalog-content\") pod \"b1650a38-f0a8-43ad-b7db-babf835c0199\" (UID: \"b1650a38-f0a8-43ad-b7db-babf835c0199\") " Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.296958 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1650a38-f0a8-43ad-b7db-babf835c0199-utilities" (OuterVolumeSpecName: "utilities") pod "b1650a38-f0a8-43ad-b7db-babf835c0199" (UID: "b1650a38-f0a8-43ad-b7db-babf835c0199"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.307664 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1650a38-f0a8-43ad-b7db-babf835c0199-kube-api-access-jp6lz" (OuterVolumeSpecName: "kube-api-access-jp6lz") pod "b1650a38-f0a8-43ad-b7db-babf835c0199" (UID: "b1650a38-f0a8-43ad-b7db-babf835c0199"). InnerVolumeSpecName "kube-api-access-jp6lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.386711 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1650a38-f0a8-43ad-b7db-babf835c0199-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1650a38-f0a8-43ad-b7db-babf835c0199" (UID: "b1650a38-f0a8-43ad-b7db-babf835c0199"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.399382 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1650a38-f0a8-43ad-b7db-babf835c0199-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.399422 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1650a38-f0a8-43ad-b7db-babf835c0199-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.399435 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jp6lz\" (UniqueName: \"kubernetes.io/projected/b1650a38-f0a8-43ad-b7db-babf835c0199-kube-api-access-jp6lz\") on node \"crc\" DevicePath \"\"" Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.473677 4771 generic.go:334] "Generic (PLEG): container finished" podID="b1650a38-f0a8-43ad-b7db-babf835c0199" containerID="dd2a4f049608bf3192fa2913c29ec7d4f18f1a732457ff7b45c3b9a06ddcdc4e" exitCode=0 Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.473733 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f847" event={"ID":"b1650a38-f0a8-43ad-b7db-babf835c0199","Type":"ContainerDied","Data":"dd2a4f049608bf3192fa2913c29ec7d4f18f1a732457ff7b45c3b9a06ddcdc4e"} Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.473770 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f847" event={"ID":"b1650a38-f0a8-43ad-b7db-babf835c0199","Type":"ContainerDied","Data":"6ce12b8321d3f5930f53ef475ea3c7567aea8d0b383c33759521968bd8265d56"} Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.473791 4771 scope.go:117] "RemoveContainer" containerID="dd2a4f049608bf3192fa2913c29ec7d4f18f1a732457ff7b45c3b9a06ddcdc4e" Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.473947 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f847" Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.508934 4771 scope.go:117] "RemoveContainer" containerID="b0df8f95cc95d07439ffa5c44dd9905e38989cef7a27a9a705aa63f2835c1dbd" Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.510037 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7f847"] Oct 02 11:05:16 crc kubenswrapper[4771]: I1002 11:05:16.521272 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7f847"] Oct 02 11:05:17 crc kubenswrapper[4771]: I1002 11:05:17.185406 4771 scope.go:117] "RemoveContainer" containerID="2659abaaf62f8054de0e9c814129ec3a7aecf1f9491abcd96afd55eba6215fe8" Oct 02 11:05:17 crc kubenswrapper[4771]: I1002 11:05:17.242487 4771 scope.go:117] "RemoveContainer" containerID="dd2a4f049608bf3192fa2913c29ec7d4f18f1a732457ff7b45c3b9a06ddcdc4e" Oct 02 11:05:17 crc kubenswrapper[4771]: E1002 11:05:17.242816 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd2a4f049608bf3192fa2913c29ec7d4f18f1a732457ff7b45c3b9a06ddcdc4e\": container with ID starting with dd2a4f049608bf3192fa2913c29ec7d4f18f1a732457ff7b45c3b9a06ddcdc4e not found: ID does not exist" containerID="dd2a4f049608bf3192fa2913c29ec7d4f18f1a732457ff7b45c3b9a06ddcdc4e" Oct 02 11:05:17 crc kubenswrapper[4771]: I1002 11:05:17.242846 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd2a4f049608bf3192fa2913c29ec7d4f18f1a732457ff7b45c3b9a06ddcdc4e"} err="failed to get container status \"dd2a4f049608bf3192fa2913c29ec7d4f18f1a732457ff7b45c3b9a06ddcdc4e\": rpc error: code = NotFound desc = could not find container \"dd2a4f049608bf3192fa2913c29ec7d4f18f1a732457ff7b45c3b9a06ddcdc4e\": container with ID starting with dd2a4f049608bf3192fa2913c29ec7d4f18f1a732457ff7b45c3b9a06ddcdc4e not found: ID does not exist" Oct 02 11:05:17 crc kubenswrapper[4771]: I1002 11:05:17.242874 4771 scope.go:117] "RemoveContainer" containerID="b0df8f95cc95d07439ffa5c44dd9905e38989cef7a27a9a705aa63f2835c1dbd" Oct 02 11:05:17 crc kubenswrapper[4771]: E1002 11:05:17.243265 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0df8f95cc95d07439ffa5c44dd9905e38989cef7a27a9a705aa63f2835c1dbd\": container with ID starting with b0df8f95cc95d07439ffa5c44dd9905e38989cef7a27a9a705aa63f2835c1dbd not found: ID does not exist" containerID="b0df8f95cc95d07439ffa5c44dd9905e38989cef7a27a9a705aa63f2835c1dbd" Oct 02 11:05:17 crc kubenswrapper[4771]: I1002 11:05:17.243289 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0df8f95cc95d07439ffa5c44dd9905e38989cef7a27a9a705aa63f2835c1dbd"} err="failed to get container status \"b0df8f95cc95d07439ffa5c44dd9905e38989cef7a27a9a705aa63f2835c1dbd\": rpc error: code = NotFound desc = could not find container \"b0df8f95cc95d07439ffa5c44dd9905e38989cef7a27a9a705aa63f2835c1dbd\": container with ID starting with b0df8f95cc95d07439ffa5c44dd9905e38989cef7a27a9a705aa63f2835c1dbd not found: ID does not exist" Oct 02 11:05:17 crc kubenswrapper[4771]: I1002 11:05:17.243302 4771 scope.go:117] "RemoveContainer" containerID="2659abaaf62f8054de0e9c814129ec3a7aecf1f9491abcd96afd55eba6215fe8" Oct 02 11:05:17 crc kubenswrapper[4771]: E1002 11:05:17.243606 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2659abaaf62f8054de0e9c814129ec3a7aecf1f9491abcd96afd55eba6215fe8\": container with ID starting with 2659abaaf62f8054de0e9c814129ec3a7aecf1f9491abcd96afd55eba6215fe8 not found: ID does not exist" containerID="2659abaaf62f8054de0e9c814129ec3a7aecf1f9491abcd96afd55eba6215fe8" Oct 02 11:05:17 crc kubenswrapper[4771]: I1002 11:05:17.243627 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2659abaaf62f8054de0e9c814129ec3a7aecf1f9491abcd96afd55eba6215fe8"} err="failed to get container status \"2659abaaf62f8054de0e9c814129ec3a7aecf1f9491abcd96afd55eba6215fe8\": rpc error: code = NotFound desc = could not find container \"2659abaaf62f8054de0e9c814129ec3a7aecf1f9491abcd96afd55eba6215fe8\": container with ID starting with 2659abaaf62f8054de0e9c814129ec3a7aecf1f9491abcd96afd55eba6215fe8 not found: ID does not exist" Oct 02 11:05:17 crc kubenswrapper[4771]: I1002 11:05:17.693842 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1650a38-f0a8-43ad-b7db-babf835c0199" path="/var/lib/kubelet/pods/b1650a38-f0a8-43ad-b7db-babf835c0199/volumes" Oct 02 11:05:21 crc kubenswrapper[4771]: E1002 11:05:21.228807 4771 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.53:50676->38.102.83.53:41045: write tcp 38.102.83.53:50676->38.102.83.53:41045: write: connection reset by peer Oct 02 11:05:28 crc kubenswrapper[4771]: I1002 11:05:28.682467 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:05:28 crc kubenswrapper[4771]: E1002 11:05:28.683243 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:05:43 crc kubenswrapper[4771]: I1002 11:05:43.692303 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:05:44 crc kubenswrapper[4771]: I1002 11:05:44.774579 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"e1807401b5f05230c75455e6d82a5a6f455c275314b1303b4322f5577e93ed5e"} Oct 02 11:08:12 crc kubenswrapper[4771]: I1002 11:08:12.145880 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:08:12 crc kubenswrapper[4771]: I1002 11:08:12.146601 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:08:42 crc kubenswrapper[4771]: I1002 11:08:42.146236 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:08:42 crc kubenswrapper[4771]: I1002 11:08:42.146984 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:09:12 crc kubenswrapper[4771]: I1002 11:09:12.146078 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:09:12 crc kubenswrapper[4771]: I1002 11:09:12.146721 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:09:12 crc kubenswrapper[4771]: I1002 11:09:12.146785 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 11:09:12 crc kubenswrapper[4771]: I1002 11:09:12.147839 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e1807401b5f05230c75455e6d82a5a6f455c275314b1303b4322f5577e93ed5e"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:09:12 crc kubenswrapper[4771]: I1002 11:09:12.147912 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://e1807401b5f05230c75455e6d82a5a6f455c275314b1303b4322f5577e93ed5e" gracePeriod=600 Oct 02 11:09:13 crc kubenswrapper[4771]: I1002 11:09:13.149914 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="e1807401b5f05230c75455e6d82a5a6f455c275314b1303b4322f5577e93ed5e" exitCode=0 Oct 02 11:09:13 crc kubenswrapper[4771]: I1002 11:09:13.149947 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"e1807401b5f05230c75455e6d82a5a6f455c275314b1303b4322f5577e93ed5e"} Oct 02 11:09:13 crc kubenswrapper[4771]: I1002 11:09:13.150543 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a"} Oct 02 11:09:13 crc kubenswrapper[4771]: I1002 11:09:13.150566 4771 scope.go:117] "RemoveContainer" containerID="95ff0027477ddbd015e77e1c51877499d14579cd3db840ffbbabd53756ae9f8b" Oct 02 11:09:38 crc kubenswrapper[4771]: I1002 11:09:38.867298 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5zqbz"] Oct 02 11:09:38 crc kubenswrapper[4771]: E1002 11:09:38.875392 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1650a38-f0a8-43ad-b7db-babf835c0199" containerName="extract-utilities" Oct 02 11:09:38 crc kubenswrapper[4771]: I1002 11:09:38.875427 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1650a38-f0a8-43ad-b7db-babf835c0199" containerName="extract-utilities" Oct 02 11:09:38 crc kubenswrapper[4771]: E1002 11:09:38.875452 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1650a38-f0a8-43ad-b7db-babf835c0199" containerName="extract-content" Oct 02 11:09:38 crc kubenswrapper[4771]: I1002 11:09:38.875458 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1650a38-f0a8-43ad-b7db-babf835c0199" containerName="extract-content" Oct 02 11:09:38 crc kubenswrapper[4771]: E1002 11:09:38.875483 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1650a38-f0a8-43ad-b7db-babf835c0199" containerName="registry-server" Oct 02 11:09:38 crc kubenswrapper[4771]: I1002 11:09:38.875488 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1650a38-f0a8-43ad-b7db-babf835c0199" containerName="registry-server" Oct 02 11:09:38 crc kubenswrapper[4771]: I1002 11:09:38.875973 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1650a38-f0a8-43ad-b7db-babf835c0199" containerName="registry-server" Oct 02 11:09:38 crc kubenswrapper[4771]: I1002 11:09:38.878201 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:38 crc kubenswrapper[4771]: I1002 11:09:38.883538 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zqbz"] Oct 02 11:09:39 crc kubenswrapper[4771]: I1002 11:09:39.018740 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2745f246-9dd5-4cd3-a383-3bbc84f649ab-catalog-content\") pod \"redhat-marketplace-5zqbz\" (UID: \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\") " pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:39 crc kubenswrapper[4771]: I1002 11:09:39.018815 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhm6r\" (UniqueName: \"kubernetes.io/projected/2745f246-9dd5-4cd3-a383-3bbc84f649ab-kube-api-access-fhm6r\") pod \"redhat-marketplace-5zqbz\" (UID: \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\") " pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:39 crc kubenswrapper[4771]: I1002 11:09:39.018970 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2745f246-9dd5-4cd3-a383-3bbc84f649ab-utilities\") pod \"redhat-marketplace-5zqbz\" (UID: \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\") " pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:39 crc kubenswrapper[4771]: I1002 11:09:39.121205 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2745f246-9dd5-4cd3-a383-3bbc84f649ab-catalog-content\") pod \"redhat-marketplace-5zqbz\" (UID: \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\") " pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:39 crc kubenswrapper[4771]: I1002 11:09:39.121275 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhm6r\" (UniqueName: \"kubernetes.io/projected/2745f246-9dd5-4cd3-a383-3bbc84f649ab-kube-api-access-fhm6r\") pod \"redhat-marketplace-5zqbz\" (UID: \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\") " pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:39 crc kubenswrapper[4771]: I1002 11:09:39.121355 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2745f246-9dd5-4cd3-a383-3bbc84f649ab-utilities\") pod \"redhat-marketplace-5zqbz\" (UID: \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\") " pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:39 crc kubenswrapper[4771]: I1002 11:09:39.121808 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2745f246-9dd5-4cd3-a383-3bbc84f649ab-catalog-content\") pod \"redhat-marketplace-5zqbz\" (UID: \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\") " pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:39 crc kubenswrapper[4771]: I1002 11:09:39.121854 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2745f246-9dd5-4cd3-a383-3bbc84f649ab-utilities\") pod \"redhat-marketplace-5zqbz\" (UID: \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\") " pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:39 crc kubenswrapper[4771]: I1002 11:09:39.143879 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhm6r\" (UniqueName: \"kubernetes.io/projected/2745f246-9dd5-4cd3-a383-3bbc84f649ab-kube-api-access-fhm6r\") pod \"redhat-marketplace-5zqbz\" (UID: \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\") " pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:39 crc kubenswrapper[4771]: I1002 11:09:39.202068 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:39 crc kubenswrapper[4771]: I1002 11:09:39.868396 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zqbz"] Oct 02 11:09:39 crc kubenswrapper[4771]: W1002 11:09:39.876556 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2745f246_9dd5_4cd3_a383_3bbc84f649ab.slice/crio-ff2afa94e12b96556680d41e4bca57dcfdb31344f0910c221ade203551dc706b WatchSource:0}: Error finding container ff2afa94e12b96556680d41e4bca57dcfdb31344f0910c221ade203551dc706b: Status 404 returned error can't find the container with id ff2afa94e12b96556680d41e4bca57dcfdb31344f0910c221ade203551dc706b Oct 02 11:09:40 crc kubenswrapper[4771]: I1002 11:09:40.479230 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zqbz" event={"ID":"2745f246-9dd5-4cd3-a383-3bbc84f649ab","Type":"ContainerStarted","Data":"ff2afa94e12b96556680d41e4bca57dcfdb31344f0910c221ade203551dc706b"} Oct 02 11:09:41 crc kubenswrapper[4771]: I1002 11:09:41.489914 4771 generic.go:334] "Generic (PLEG): container finished" podID="2745f246-9dd5-4cd3-a383-3bbc84f649ab" containerID="b691f315f865d0eebe785e81ebccb994a60762a7050a163b11cbaaaae16dd0a4" exitCode=0 Oct 02 11:09:41 crc kubenswrapper[4771]: I1002 11:09:41.489974 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zqbz" event={"ID":"2745f246-9dd5-4cd3-a383-3bbc84f649ab","Type":"ContainerDied","Data":"b691f315f865d0eebe785e81ebccb994a60762a7050a163b11cbaaaae16dd0a4"} Oct 02 11:09:43 crc kubenswrapper[4771]: I1002 11:09:43.512745 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zqbz" event={"ID":"2745f246-9dd5-4cd3-a383-3bbc84f649ab","Type":"ContainerStarted","Data":"c3c5df9dec206b43edf29e60c16c9dd4f55138fc27a839ce95dd553e894a67ec"} Oct 02 11:09:44 crc kubenswrapper[4771]: I1002 11:09:44.528485 4771 generic.go:334] "Generic (PLEG): container finished" podID="2745f246-9dd5-4cd3-a383-3bbc84f649ab" containerID="c3c5df9dec206b43edf29e60c16c9dd4f55138fc27a839ce95dd553e894a67ec" exitCode=0 Oct 02 11:09:44 crc kubenswrapper[4771]: I1002 11:09:44.528551 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zqbz" event={"ID":"2745f246-9dd5-4cd3-a383-3bbc84f649ab","Type":"ContainerDied","Data":"c3c5df9dec206b43edf29e60c16c9dd4f55138fc27a839ce95dd553e894a67ec"} Oct 02 11:09:45 crc kubenswrapper[4771]: I1002 11:09:45.542815 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zqbz" event={"ID":"2745f246-9dd5-4cd3-a383-3bbc84f649ab","Type":"ContainerStarted","Data":"b166dbf640cb510177bf6dbeb62417f8256403f15b0bdafecd17a9830a9c260c"} Oct 02 11:09:45 crc kubenswrapper[4771]: I1002 11:09:45.579856 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5zqbz" podStartSLOduration=3.991664117 podStartE2EDuration="7.5798187s" podCreationTimestamp="2025-10-02 11:09:38 +0000 UTC" firstStartedPulling="2025-10-02 11:09:41.491881126 +0000 UTC m=+5569.139566193" lastFinishedPulling="2025-10-02 11:09:45.080035709 +0000 UTC m=+5572.727720776" observedRunningTime="2025-10-02 11:09:45.560591761 +0000 UTC m=+5573.208276828" watchObservedRunningTime="2025-10-02 11:09:45.5798187 +0000 UTC m=+5573.227503767" Oct 02 11:09:49 crc kubenswrapper[4771]: I1002 11:09:49.203493 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:49 crc kubenswrapper[4771]: I1002 11:09:49.204049 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:49 crc kubenswrapper[4771]: I1002 11:09:49.263045 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:59 crc kubenswrapper[4771]: I1002 11:09:59.251544 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:09:59 crc kubenswrapper[4771]: I1002 11:09:59.327283 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zqbz"] Oct 02 11:09:59 crc kubenswrapper[4771]: I1002 11:09:59.723394 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5zqbz" podUID="2745f246-9dd5-4cd3-a383-3bbc84f649ab" containerName="registry-server" containerID="cri-o://b166dbf640cb510177bf6dbeb62417f8256403f15b0bdafecd17a9830a9c260c" gracePeriod=2 Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.275286 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.343016 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2745f246-9dd5-4cd3-a383-3bbc84f649ab-catalog-content\") pod \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\" (UID: \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\") " Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.344290 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhm6r\" (UniqueName: \"kubernetes.io/projected/2745f246-9dd5-4cd3-a383-3bbc84f649ab-kube-api-access-fhm6r\") pod \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\" (UID: \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\") " Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.344452 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2745f246-9dd5-4cd3-a383-3bbc84f649ab-utilities\") pod \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\" (UID: \"2745f246-9dd5-4cd3-a383-3bbc84f649ab\") " Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.346575 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2745f246-9dd5-4cd3-a383-3bbc84f649ab-utilities" (OuterVolumeSpecName: "utilities") pod "2745f246-9dd5-4cd3-a383-3bbc84f649ab" (UID: "2745f246-9dd5-4cd3-a383-3bbc84f649ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.354702 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2745f246-9dd5-4cd3-a383-3bbc84f649ab-kube-api-access-fhm6r" (OuterVolumeSpecName: "kube-api-access-fhm6r") pod "2745f246-9dd5-4cd3-a383-3bbc84f649ab" (UID: "2745f246-9dd5-4cd3-a383-3bbc84f649ab"). InnerVolumeSpecName "kube-api-access-fhm6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.359142 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2745f246-9dd5-4cd3-a383-3bbc84f649ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2745f246-9dd5-4cd3-a383-3bbc84f649ab" (UID: "2745f246-9dd5-4cd3-a383-3bbc84f649ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.447809 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhm6r\" (UniqueName: \"kubernetes.io/projected/2745f246-9dd5-4cd3-a383-3bbc84f649ab-kube-api-access-fhm6r\") on node \"crc\" DevicePath \"\"" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.447904 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2745f246-9dd5-4cd3-a383-3bbc84f649ab-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.447918 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2745f246-9dd5-4cd3-a383-3bbc84f649ab-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.736348 4771 generic.go:334] "Generic (PLEG): container finished" podID="2745f246-9dd5-4cd3-a383-3bbc84f649ab" containerID="b166dbf640cb510177bf6dbeb62417f8256403f15b0bdafecd17a9830a9c260c" exitCode=0 Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.736405 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zqbz" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.736432 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zqbz" event={"ID":"2745f246-9dd5-4cd3-a383-3bbc84f649ab","Type":"ContainerDied","Data":"b166dbf640cb510177bf6dbeb62417f8256403f15b0bdafecd17a9830a9c260c"} Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.737080 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zqbz" event={"ID":"2745f246-9dd5-4cd3-a383-3bbc84f649ab","Type":"ContainerDied","Data":"ff2afa94e12b96556680d41e4bca57dcfdb31344f0910c221ade203551dc706b"} Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.737103 4771 scope.go:117] "RemoveContainer" containerID="b166dbf640cb510177bf6dbeb62417f8256403f15b0bdafecd17a9830a9c260c" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.768648 4771 scope.go:117] "RemoveContainer" containerID="c3c5df9dec206b43edf29e60c16c9dd4f55138fc27a839ce95dd553e894a67ec" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.771801 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zqbz"] Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.781721 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zqbz"] Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.790468 4771 scope.go:117] "RemoveContainer" containerID="b691f315f865d0eebe785e81ebccb994a60762a7050a163b11cbaaaae16dd0a4" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.847496 4771 scope.go:117] "RemoveContainer" containerID="b166dbf640cb510177bf6dbeb62417f8256403f15b0bdafecd17a9830a9c260c" Oct 02 11:10:00 crc kubenswrapper[4771]: E1002 11:10:00.848400 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b166dbf640cb510177bf6dbeb62417f8256403f15b0bdafecd17a9830a9c260c\": container with ID starting with b166dbf640cb510177bf6dbeb62417f8256403f15b0bdafecd17a9830a9c260c not found: ID does not exist" containerID="b166dbf640cb510177bf6dbeb62417f8256403f15b0bdafecd17a9830a9c260c" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.848574 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b166dbf640cb510177bf6dbeb62417f8256403f15b0bdafecd17a9830a9c260c"} err="failed to get container status \"b166dbf640cb510177bf6dbeb62417f8256403f15b0bdafecd17a9830a9c260c\": rpc error: code = NotFound desc = could not find container \"b166dbf640cb510177bf6dbeb62417f8256403f15b0bdafecd17a9830a9c260c\": container with ID starting with b166dbf640cb510177bf6dbeb62417f8256403f15b0bdafecd17a9830a9c260c not found: ID does not exist" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.848725 4771 scope.go:117] "RemoveContainer" containerID="c3c5df9dec206b43edf29e60c16c9dd4f55138fc27a839ce95dd553e894a67ec" Oct 02 11:10:00 crc kubenswrapper[4771]: E1002 11:10:00.849314 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3c5df9dec206b43edf29e60c16c9dd4f55138fc27a839ce95dd553e894a67ec\": container with ID starting with c3c5df9dec206b43edf29e60c16c9dd4f55138fc27a839ce95dd553e894a67ec not found: ID does not exist" containerID="c3c5df9dec206b43edf29e60c16c9dd4f55138fc27a839ce95dd553e894a67ec" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.849362 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3c5df9dec206b43edf29e60c16c9dd4f55138fc27a839ce95dd553e894a67ec"} err="failed to get container status \"c3c5df9dec206b43edf29e60c16c9dd4f55138fc27a839ce95dd553e894a67ec\": rpc error: code = NotFound desc = could not find container \"c3c5df9dec206b43edf29e60c16c9dd4f55138fc27a839ce95dd553e894a67ec\": container with ID starting with c3c5df9dec206b43edf29e60c16c9dd4f55138fc27a839ce95dd553e894a67ec not found: ID does not exist" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.849391 4771 scope.go:117] "RemoveContainer" containerID="b691f315f865d0eebe785e81ebccb994a60762a7050a163b11cbaaaae16dd0a4" Oct 02 11:10:00 crc kubenswrapper[4771]: E1002 11:10:00.849728 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b691f315f865d0eebe785e81ebccb994a60762a7050a163b11cbaaaae16dd0a4\": container with ID starting with b691f315f865d0eebe785e81ebccb994a60762a7050a163b11cbaaaae16dd0a4 not found: ID does not exist" containerID="b691f315f865d0eebe785e81ebccb994a60762a7050a163b11cbaaaae16dd0a4" Oct 02 11:10:00 crc kubenswrapper[4771]: I1002 11:10:00.849761 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b691f315f865d0eebe785e81ebccb994a60762a7050a163b11cbaaaae16dd0a4"} err="failed to get container status \"b691f315f865d0eebe785e81ebccb994a60762a7050a163b11cbaaaae16dd0a4\": rpc error: code = NotFound desc = could not find container \"b691f315f865d0eebe785e81ebccb994a60762a7050a163b11cbaaaae16dd0a4\": container with ID starting with b691f315f865d0eebe785e81ebccb994a60762a7050a163b11cbaaaae16dd0a4 not found: ID does not exist" Oct 02 11:10:01 crc kubenswrapper[4771]: I1002 11:10:01.712619 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2745f246-9dd5-4cd3-a383-3bbc84f649ab" path="/var/lib/kubelet/pods/2745f246-9dd5-4cd3-a383-3bbc84f649ab/volumes" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.383966 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zjwxp"] Oct 02 11:10:54 crc kubenswrapper[4771]: E1002 11:10:54.384947 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2745f246-9dd5-4cd3-a383-3bbc84f649ab" containerName="extract-utilities" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.384960 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2745f246-9dd5-4cd3-a383-3bbc84f649ab" containerName="extract-utilities" Oct 02 11:10:54 crc kubenswrapper[4771]: E1002 11:10:54.384982 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2745f246-9dd5-4cd3-a383-3bbc84f649ab" containerName="extract-content" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.384988 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2745f246-9dd5-4cd3-a383-3bbc84f649ab" containerName="extract-content" Oct 02 11:10:54 crc kubenswrapper[4771]: E1002 11:10:54.385012 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2745f246-9dd5-4cd3-a383-3bbc84f649ab" containerName="registry-server" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.385019 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="2745f246-9dd5-4cd3-a383-3bbc84f649ab" containerName="registry-server" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.385245 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="2745f246-9dd5-4cd3-a383-3bbc84f649ab" containerName="registry-server" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.387004 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.401733 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zjwxp"] Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.466177 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c39c79fc-d3d3-458b-be40-45c01ac10245-utilities\") pod \"certified-operators-zjwxp\" (UID: \"c39c79fc-d3d3-458b-be40-45c01ac10245\") " pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.466294 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpmdb\" (UniqueName: \"kubernetes.io/projected/c39c79fc-d3d3-458b-be40-45c01ac10245-kube-api-access-mpmdb\") pod \"certified-operators-zjwxp\" (UID: \"c39c79fc-d3d3-458b-be40-45c01ac10245\") " pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.466340 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c39c79fc-d3d3-458b-be40-45c01ac10245-catalog-content\") pod \"certified-operators-zjwxp\" (UID: \"c39c79fc-d3d3-458b-be40-45c01ac10245\") " pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.568243 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c39c79fc-d3d3-458b-be40-45c01ac10245-utilities\") pod \"certified-operators-zjwxp\" (UID: \"c39c79fc-d3d3-458b-be40-45c01ac10245\") " pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.568436 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpmdb\" (UniqueName: \"kubernetes.io/projected/c39c79fc-d3d3-458b-be40-45c01ac10245-kube-api-access-mpmdb\") pod \"certified-operators-zjwxp\" (UID: \"c39c79fc-d3d3-458b-be40-45c01ac10245\") " pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.568498 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c39c79fc-d3d3-458b-be40-45c01ac10245-catalog-content\") pod \"certified-operators-zjwxp\" (UID: \"c39c79fc-d3d3-458b-be40-45c01ac10245\") " pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.568683 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c39c79fc-d3d3-458b-be40-45c01ac10245-utilities\") pod \"certified-operators-zjwxp\" (UID: \"c39c79fc-d3d3-458b-be40-45c01ac10245\") " pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.568865 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c39c79fc-d3d3-458b-be40-45c01ac10245-catalog-content\") pod \"certified-operators-zjwxp\" (UID: \"c39c79fc-d3d3-458b-be40-45c01ac10245\") " pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.586982 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpmdb\" (UniqueName: \"kubernetes.io/projected/c39c79fc-d3d3-458b-be40-45c01ac10245-kube-api-access-mpmdb\") pod \"certified-operators-zjwxp\" (UID: \"c39c79fc-d3d3-458b-be40-45c01ac10245\") " pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:10:54 crc kubenswrapper[4771]: I1002 11:10:54.733770 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:10:55 crc kubenswrapper[4771]: I1002 11:10:55.342420 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zjwxp"] Oct 02 11:10:56 crc kubenswrapper[4771]: I1002 11:10:56.321510 4771 generic.go:334] "Generic (PLEG): container finished" podID="c39c79fc-d3d3-458b-be40-45c01ac10245" containerID="fc080978c1d045a5f4acac1854c11606c20535d597acd28409e8ab4e76cfba19" exitCode=0 Oct 02 11:10:56 crc kubenswrapper[4771]: I1002 11:10:56.321634 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjwxp" event={"ID":"c39c79fc-d3d3-458b-be40-45c01ac10245","Type":"ContainerDied","Data":"fc080978c1d045a5f4acac1854c11606c20535d597acd28409e8ab4e76cfba19"} Oct 02 11:10:56 crc kubenswrapper[4771]: I1002 11:10:56.322943 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjwxp" event={"ID":"c39c79fc-d3d3-458b-be40-45c01ac10245","Type":"ContainerStarted","Data":"ab1cf02080f099fdaefed3df4a3fac0a3fbc1250c961f0c4dfc96f6052de59d0"} Oct 02 11:10:56 crc kubenswrapper[4771]: I1002 11:10:56.323837 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:10:57 crc kubenswrapper[4771]: I1002 11:10:57.335704 4771 generic.go:334] "Generic (PLEG): container finished" podID="c39c79fc-d3d3-458b-be40-45c01ac10245" containerID="1662ea18c4856a18cffbb68d17c0b78e22f6ca2aa561a645c4fbdd4b365ecb08" exitCode=0 Oct 02 11:10:57 crc kubenswrapper[4771]: I1002 11:10:57.335817 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjwxp" event={"ID":"c39c79fc-d3d3-458b-be40-45c01ac10245","Type":"ContainerDied","Data":"1662ea18c4856a18cffbb68d17c0b78e22f6ca2aa561a645c4fbdd4b365ecb08"} Oct 02 11:10:58 crc kubenswrapper[4771]: I1002 11:10:58.353763 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjwxp" event={"ID":"c39c79fc-d3d3-458b-be40-45c01ac10245","Type":"ContainerStarted","Data":"b5449d60bccbf504f24b6aed7f3afdf0b8295a36793f205feb7ed3bc2e39301d"} Oct 02 11:10:58 crc kubenswrapper[4771]: I1002 11:10:58.382795 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zjwxp" podStartSLOduration=2.86907503 podStartE2EDuration="4.382771905s" podCreationTimestamp="2025-10-02 11:10:54 +0000 UTC" firstStartedPulling="2025-10-02 11:10:56.323549892 +0000 UTC m=+5643.971234959" lastFinishedPulling="2025-10-02 11:10:57.837246767 +0000 UTC m=+5645.484931834" observedRunningTime="2025-10-02 11:10:58.376480851 +0000 UTC m=+5646.024165908" watchObservedRunningTime="2025-10-02 11:10:58.382771905 +0000 UTC m=+5646.030456972" Oct 02 11:11:04 crc kubenswrapper[4771]: I1002 11:11:04.734749 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:11:04 crc kubenswrapper[4771]: I1002 11:11:04.735604 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:11:04 crc kubenswrapper[4771]: I1002 11:11:04.815869 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:11:05 crc kubenswrapper[4771]: I1002 11:11:05.488366 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:11:05 crc kubenswrapper[4771]: I1002 11:11:05.545981 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zjwxp"] Oct 02 11:11:07 crc kubenswrapper[4771]: I1002 11:11:07.460395 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zjwxp" podUID="c39c79fc-d3d3-458b-be40-45c01ac10245" containerName="registry-server" containerID="cri-o://b5449d60bccbf504f24b6aed7f3afdf0b8295a36793f205feb7ed3bc2e39301d" gracePeriod=2 Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.028812 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.111963 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c39c79fc-d3d3-458b-be40-45c01ac10245-catalog-content\") pod \"c39c79fc-d3d3-458b-be40-45c01ac10245\" (UID: \"c39c79fc-d3d3-458b-be40-45c01ac10245\") " Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.112164 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpmdb\" (UniqueName: \"kubernetes.io/projected/c39c79fc-d3d3-458b-be40-45c01ac10245-kube-api-access-mpmdb\") pod \"c39c79fc-d3d3-458b-be40-45c01ac10245\" (UID: \"c39c79fc-d3d3-458b-be40-45c01ac10245\") " Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.112225 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c39c79fc-d3d3-458b-be40-45c01ac10245-utilities\") pod \"c39c79fc-d3d3-458b-be40-45c01ac10245\" (UID: \"c39c79fc-d3d3-458b-be40-45c01ac10245\") " Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.113771 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c39c79fc-d3d3-458b-be40-45c01ac10245-utilities" (OuterVolumeSpecName: "utilities") pod "c39c79fc-d3d3-458b-be40-45c01ac10245" (UID: "c39c79fc-d3d3-458b-be40-45c01ac10245"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.125280 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c39c79fc-d3d3-458b-be40-45c01ac10245-kube-api-access-mpmdb" (OuterVolumeSpecName: "kube-api-access-mpmdb") pod "c39c79fc-d3d3-458b-be40-45c01ac10245" (UID: "c39c79fc-d3d3-458b-be40-45c01ac10245"). InnerVolumeSpecName "kube-api-access-mpmdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.156821 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c39c79fc-d3d3-458b-be40-45c01ac10245-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c39c79fc-d3d3-458b-be40-45c01ac10245" (UID: "c39c79fc-d3d3-458b-be40-45c01ac10245"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.218284 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c39c79fc-d3d3-458b-be40-45c01ac10245-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.218328 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c39c79fc-d3d3-458b-be40-45c01ac10245-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.218343 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpmdb\" (UniqueName: \"kubernetes.io/projected/c39c79fc-d3d3-458b-be40-45c01ac10245-kube-api-access-mpmdb\") on node \"crc\" DevicePath \"\"" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.478913 4771 generic.go:334] "Generic (PLEG): container finished" podID="c39c79fc-d3d3-458b-be40-45c01ac10245" containerID="b5449d60bccbf504f24b6aed7f3afdf0b8295a36793f205feb7ed3bc2e39301d" exitCode=0 Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.478979 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjwxp" event={"ID":"c39c79fc-d3d3-458b-be40-45c01ac10245","Type":"ContainerDied","Data":"b5449d60bccbf504f24b6aed7f3afdf0b8295a36793f205feb7ed3bc2e39301d"} Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.479069 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjwxp" event={"ID":"c39c79fc-d3d3-458b-be40-45c01ac10245","Type":"ContainerDied","Data":"ab1cf02080f099fdaefed3df4a3fac0a3fbc1250c961f0c4dfc96f6052de59d0"} Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.479110 4771 scope.go:117] "RemoveContainer" containerID="b5449d60bccbf504f24b6aed7f3afdf0b8295a36793f205feb7ed3bc2e39301d" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.481061 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zjwxp" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.518742 4771 scope.go:117] "RemoveContainer" containerID="1662ea18c4856a18cffbb68d17c0b78e22f6ca2aa561a645c4fbdd4b365ecb08" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.531408 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zjwxp"] Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.545893 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zjwxp"] Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.557994 4771 scope.go:117] "RemoveContainer" containerID="fc080978c1d045a5f4acac1854c11606c20535d597acd28409e8ab4e76cfba19" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.605001 4771 scope.go:117] "RemoveContainer" containerID="b5449d60bccbf504f24b6aed7f3afdf0b8295a36793f205feb7ed3bc2e39301d" Oct 02 11:11:08 crc kubenswrapper[4771]: E1002 11:11:08.605548 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5449d60bccbf504f24b6aed7f3afdf0b8295a36793f205feb7ed3bc2e39301d\": container with ID starting with b5449d60bccbf504f24b6aed7f3afdf0b8295a36793f205feb7ed3bc2e39301d not found: ID does not exist" containerID="b5449d60bccbf504f24b6aed7f3afdf0b8295a36793f205feb7ed3bc2e39301d" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.605577 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5449d60bccbf504f24b6aed7f3afdf0b8295a36793f205feb7ed3bc2e39301d"} err="failed to get container status \"b5449d60bccbf504f24b6aed7f3afdf0b8295a36793f205feb7ed3bc2e39301d\": rpc error: code = NotFound desc = could not find container \"b5449d60bccbf504f24b6aed7f3afdf0b8295a36793f205feb7ed3bc2e39301d\": container with ID starting with b5449d60bccbf504f24b6aed7f3afdf0b8295a36793f205feb7ed3bc2e39301d not found: ID does not exist" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.605597 4771 scope.go:117] "RemoveContainer" containerID="1662ea18c4856a18cffbb68d17c0b78e22f6ca2aa561a645c4fbdd4b365ecb08" Oct 02 11:11:08 crc kubenswrapper[4771]: E1002 11:11:08.606041 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1662ea18c4856a18cffbb68d17c0b78e22f6ca2aa561a645c4fbdd4b365ecb08\": container with ID starting with 1662ea18c4856a18cffbb68d17c0b78e22f6ca2aa561a645c4fbdd4b365ecb08 not found: ID does not exist" containerID="1662ea18c4856a18cffbb68d17c0b78e22f6ca2aa561a645c4fbdd4b365ecb08" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.606098 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1662ea18c4856a18cffbb68d17c0b78e22f6ca2aa561a645c4fbdd4b365ecb08"} err="failed to get container status \"1662ea18c4856a18cffbb68d17c0b78e22f6ca2aa561a645c4fbdd4b365ecb08\": rpc error: code = NotFound desc = could not find container \"1662ea18c4856a18cffbb68d17c0b78e22f6ca2aa561a645c4fbdd4b365ecb08\": container with ID starting with 1662ea18c4856a18cffbb68d17c0b78e22f6ca2aa561a645c4fbdd4b365ecb08 not found: ID does not exist" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.606163 4771 scope.go:117] "RemoveContainer" containerID="fc080978c1d045a5f4acac1854c11606c20535d597acd28409e8ab4e76cfba19" Oct 02 11:11:08 crc kubenswrapper[4771]: E1002 11:11:08.606797 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc080978c1d045a5f4acac1854c11606c20535d597acd28409e8ab4e76cfba19\": container with ID starting with fc080978c1d045a5f4acac1854c11606c20535d597acd28409e8ab4e76cfba19 not found: ID does not exist" containerID="fc080978c1d045a5f4acac1854c11606c20535d597acd28409e8ab4e76cfba19" Oct 02 11:11:08 crc kubenswrapper[4771]: I1002 11:11:08.606845 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc080978c1d045a5f4acac1854c11606c20535d597acd28409e8ab4e76cfba19"} err="failed to get container status \"fc080978c1d045a5f4acac1854c11606c20535d597acd28409e8ab4e76cfba19\": rpc error: code = NotFound desc = could not find container \"fc080978c1d045a5f4acac1854c11606c20535d597acd28409e8ab4e76cfba19\": container with ID starting with fc080978c1d045a5f4acac1854c11606c20535d597acd28409e8ab4e76cfba19 not found: ID does not exist" Oct 02 11:11:09 crc kubenswrapper[4771]: I1002 11:11:09.695443 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c39c79fc-d3d3-458b-be40-45c01ac10245" path="/var/lib/kubelet/pods/c39c79fc-d3d3-458b-be40-45c01ac10245/volumes" Oct 02 11:11:12 crc kubenswrapper[4771]: I1002 11:11:12.146237 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:11:12 crc kubenswrapper[4771]: I1002 11:11:12.146795 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:11:42 crc kubenswrapper[4771]: I1002 11:11:42.146457 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:11:42 crc kubenswrapper[4771]: I1002 11:11:42.147896 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:12:12 crc kubenswrapper[4771]: I1002 11:12:12.145977 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:12:12 crc kubenswrapper[4771]: I1002 11:12:12.147601 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:12:12 crc kubenswrapper[4771]: I1002 11:12:12.147822 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 11:12:12 crc kubenswrapper[4771]: I1002 11:12:12.149602 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:12:12 crc kubenswrapper[4771]: I1002 11:12:12.149746 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" gracePeriod=600 Oct 02 11:12:12 crc kubenswrapper[4771]: E1002 11:12:12.278443 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:12:13 crc kubenswrapper[4771]: I1002 11:12:13.216724 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" exitCode=0 Oct 02 11:12:13 crc kubenswrapper[4771]: I1002 11:12:13.216791 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a"} Oct 02 11:12:13 crc kubenswrapper[4771]: I1002 11:12:13.216990 4771 scope.go:117] "RemoveContainer" containerID="e1807401b5f05230c75455e6d82a5a6f455c275314b1303b4322f5577e93ed5e" Oct 02 11:12:13 crc kubenswrapper[4771]: I1002 11:12:13.217665 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:12:13 crc kubenswrapper[4771]: E1002 11:12:13.218010 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:12:25 crc kubenswrapper[4771]: I1002 11:12:25.681899 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:12:25 crc kubenswrapper[4771]: E1002 11:12:25.682923 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:12:39 crc kubenswrapper[4771]: I1002 11:12:39.682257 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:12:39 crc kubenswrapper[4771]: E1002 11:12:39.686777 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:12:52 crc kubenswrapper[4771]: I1002 11:12:52.681380 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:12:52 crc kubenswrapper[4771]: E1002 11:12:52.682039 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:13:07 crc kubenswrapper[4771]: I1002 11:13:07.681385 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:13:07 crc kubenswrapper[4771]: E1002 11:13:07.682259 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:13:18 crc kubenswrapper[4771]: I1002 11:13:18.681830 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:13:18 crc kubenswrapper[4771]: E1002 11:13:18.682999 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:13:29 crc kubenswrapper[4771]: I1002 11:13:29.681000 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:13:29 crc kubenswrapper[4771]: E1002 11:13:29.681913 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:13:41 crc kubenswrapper[4771]: I1002 11:13:41.683384 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:13:41 crc kubenswrapper[4771]: E1002 11:13:41.684797 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:13:53 crc kubenswrapper[4771]: I1002 11:13:53.690643 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:13:53 crc kubenswrapper[4771]: E1002 11:13:53.691600 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:14:05 crc kubenswrapper[4771]: I1002 11:14:05.683166 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:14:05 crc kubenswrapper[4771]: E1002 11:14:05.685001 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:14:18 crc kubenswrapper[4771]: I1002 11:14:18.683393 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:14:18 crc kubenswrapper[4771]: E1002 11:14:18.684473 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:14:30 crc kubenswrapper[4771]: I1002 11:14:30.681215 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:14:30 crc kubenswrapper[4771]: E1002 11:14:30.681854 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:14:42 crc kubenswrapper[4771]: I1002 11:14:42.682980 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:14:42 crc kubenswrapper[4771]: E1002 11:14:42.683737 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:14:57 crc kubenswrapper[4771]: I1002 11:14:57.682026 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:14:57 crc kubenswrapper[4771]: E1002 11:14:57.683543 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.183506 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw"] Oct 02 11:15:00 crc kubenswrapper[4771]: E1002 11:15:00.184654 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c39c79fc-d3d3-458b-be40-45c01ac10245" containerName="extract-content" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.184669 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="c39c79fc-d3d3-458b-be40-45c01ac10245" containerName="extract-content" Oct 02 11:15:00 crc kubenswrapper[4771]: E1002 11:15:00.184680 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c39c79fc-d3d3-458b-be40-45c01ac10245" containerName="extract-utilities" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.184687 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="c39c79fc-d3d3-458b-be40-45c01ac10245" containerName="extract-utilities" Oct 02 11:15:00 crc kubenswrapper[4771]: E1002 11:15:00.184734 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c39c79fc-d3d3-458b-be40-45c01ac10245" containerName="registry-server" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.184741 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="c39c79fc-d3d3-458b-be40-45c01ac10245" containerName="registry-server" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.184963 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="c39c79fc-d3d3-458b-be40-45c01ac10245" containerName="registry-server" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.185829 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.188444 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.188629 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.196351 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw"] Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.342311 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36b7c50d-0488-4ad1-b2e5-756416f88ca6-config-volume\") pod \"collect-profiles-29323395-xmfpw\" (UID: \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.342565 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sffbd\" (UniqueName: \"kubernetes.io/projected/36b7c50d-0488-4ad1-b2e5-756416f88ca6-kube-api-access-sffbd\") pod \"collect-profiles-29323395-xmfpw\" (UID: \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.342653 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36b7c50d-0488-4ad1-b2e5-756416f88ca6-secret-volume\") pod \"collect-profiles-29323395-xmfpw\" (UID: \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.445007 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36b7c50d-0488-4ad1-b2e5-756416f88ca6-config-volume\") pod \"collect-profiles-29323395-xmfpw\" (UID: \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.445173 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sffbd\" (UniqueName: \"kubernetes.io/projected/36b7c50d-0488-4ad1-b2e5-756416f88ca6-kube-api-access-sffbd\") pod \"collect-profiles-29323395-xmfpw\" (UID: \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.445213 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36b7c50d-0488-4ad1-b2e5-756416f88ca6-secret-volume\") pod \"collect-profiles-29323395-xmfpw\" (UID: \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.446072 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36b7c50d-0488-4ad1-b2e5-756416f88ca6-config-volume\") pod \"collect-profiles-29323395-xmfpw\" (UID: \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.451230 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36b7c50d-0488-4ad1-b2e5-756416f88ca6-secret-volume\") pod \"collect-profiles-29323395-xmfpw\" (UID: \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.461809 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sffbd\" (UniqueName: \"kubernetes.io/projected/36b7c50d-0488-4ad1-b2e5-756416f88ca6-kube-api-access-sffbd\") pod \"collect-profiles-29323395-xmfpw\" (UID: \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" Oct 02 11:15:00 crc kubenswrapper[4771]: I1002 11:15:00.536404 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" Oct 02 11:15:01 crc kubenswrapper[4771]: I1002 11:15:01.020812 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw"] Oct 02 11:15:01 crc kubenswrapper[4771]: I1002 11:15:01.067796 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" event={"ID":"36b7c50d-0488-4ad1-b2e5-756416f88ca6","Type":"ContainerStarted","Data":"22b00e788aad7554b0b60b3816e26ca04507e1bdda01d653bcfb1e163b1676b6"} Oct 02 11:15:02 crc kubenswrapper[4771]: I1002 11:15:02.078712 4771 generic.go:334] "Generic (PLEG): container finished" podID="36b7c50d-0488-4ad1-b2e5-756416f88ca6" containerID="112057a734f698c94d1172b43104c19bfa1bbfa09851d4cc26ca3dda69b2b7e2" exitCode=0 Oct 02 11:15:02 crc kubenswrapper[4771]: I1002 11:15:02.078793 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" event={"ID":"36b7c50d-0488-4ad1-b2e5-756416f88ca6","Type":"ContainerDied","Data":"112057a734f698c94d1172b43104c19bfa1bbfa09851d4cc26ca3dda69b2b7e2"} Oct 02 11:15:03 crc kubenswrapper[4771]: I1002 11:15:03.620885 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" Oct 02 11:15:03 crc kubenswrapper[4771]: I1002 11:15:03.722345 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36b7c50d-0488-4ad1-b2e5-756416f88ca6-config-volume\") pod \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\" (UID: \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\") " Oct 02 11:15:03 crc kubenswrapper[4771]: I1002 11:15:03.722464 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36b7c50d-0488-4ad1-b2e5-756416f88ca6-secret-volume\") pod \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\" (UID: \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\") " Oct 02 11:15:03 crc kubenswrapper[4771]: I1002 11:15:03.722663 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sffbd\" (UniqueName: \"kubernetes.io/projected/36b7c50d-0488-4ad1-b2e5-756416f88ca6-kube-api-access-sffbd\") pod \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\" (UID: \"36b7c50d-0488-4ad1-b2e5-756416f88ca6\") " Oct 02 11:15:03 crc kubenswrapper[4771]: I1002 11:15:03.723463 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36b7c50d-0488-4ad1-b2e5-756416f88ca6-config-volume" (OuterVolumeSpecName: "config-volume") pod "36b7c50d-0488-4ad1-b2e5-756416f88ca6" (UID: "36b7c50d-0488-4ad1-b2e5-756416f88ca6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:03 crc kubenswrapper[4771]: I1002 11:15:03.729007 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36b7c50d-0488-4ad1-b2e5-756416f88ca6-kube-api-access-sffbd" (OuterVolumeSpecName: "kube-api-access-sffbd") pod "36b7c50d-0488-4ad1-b2e5-756416f88ca6" (UID: "36b7c50d-0488-4ad1-b2e5-756416f88ca6"). InnerVolumeSpecName "kube-api-access-sffbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:03 crc kubenswrapper[4771]: I1002 11:15:03.729568 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36b7c50d-0488-4ad1-b2e5-756416f88ca6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "36b7c50d-0488-4ad1-b2e5-756416f88ca6" (UID: "36b7c50d-0488-4ad1-b2e5-756416f88ca6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:03 crc kubenswrapper[4771]: I1002 11:15:03.825575 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sffbd\" (UniqueName: \"kubernetes.io/projected/36b7c50d-0488-4ad1-b2e5-756416f88ca6-kube-api-access-sffbd\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:03 crc kubenswrapper[4771]: I1002 11:15:03.825612 4771 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36b7c50d-0488-4ad1-b2e5-756416f88ca6-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:03 crc kubenswrapper[4771]: I1002 11:15:03.825621 4771 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/36b7c50d-0488-4ad1-b2e5-756416f88ca6-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:04 crc kubenswrapper[4771]: I1002 11:15:04.105824 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" event={"ID":"36b7c50d-0488-4ad1-b2e5-756416f88ca6","Type":"ContainerDied","Data":"22b00e788aad7554b0b60b3816e26ca04507e1bdda01d653bcfb1e163b1676b6"} Oct 02 11:15:04 crc kubenswrapper[4771]: I1002 11:15:04.105901 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22b00e788aad7554b0b60b3816e26ca04507e1bdda01d653bcfb1e163b1676b6" Oct 02 11:15:04 crc kubenswrapper[4771]: I1002 11:15:04.105870 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw" Oct 02 11:15:04 crc kubenswrapper[4771]: I1002 11:15:04.699761 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898"] Oct 02 11:15:04 crc kubenswrapper[4771]: I1002 11:15:04.711299 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323350-7j898"] Oct 02 11:15:05 crc kubenswrapper[4771]: I1002 11:15:05.741480 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15d11e44-5df9-4282-9bd7-b3554aa7b26a" path="/var/lib/kubelet/pods/15d11e44-5df9-4282-9bd7-b3554aa7b26a/volumes" Oct 02 11:15:06 crc kubenswrapper[4771]: I1002 11:15:06.563071 4771 scope.go:117] "RemoveContainer" containerID="82601b0ebfffbc20917c1dc42d76d82c0dfdd4c4e401ec440e0066b50b88c533" Oct 02 11:15:12 crc kubenswrapper[4771]: I1002 11:15:12.681154 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:15:12 crc kubenswrapper[4771]: E1002 11:15:12.682115 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:15:23 crc kubenswrapper[4771]: I1002 11:15:23.700798 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:15:23 crc kubenswrapper[4771]: E1002 11:15:23.703737 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:15:34 crc kubenswrapper[4771]: I1002 11:15:34.681223 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:15:34 crc kubenswrapper[4771]: E1002 11:15:34.682239 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:15:45 crc kubenswrapper[4771]: I1002 11:15:45.682275 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:15:45 crc kubenswrapper[4771]: E1002 11:15:45.683182 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.521775 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rcqmp"] Oct 02 11:15:49 crc kubenswrapper[4771]: E1002 11:15:49.523181 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b7c50d-0488-4ad1-b2e5-756416f88ca6" containerName="collect-profiles" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.523199 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b7c50d-0488-4ad1-b2e5-756416f88ca6" containerName="collect-profiles" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.523484 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="36b7c50d-0488-4ad1-b2e5-756416f88ca6" containerName="collect-profiles" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.525808 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.540587 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rcqmp"] Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.669381 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/323d9415-907b-4449-867a-2fb4659c9881-catalog-content\") pod \"redhat-operators-rcqmp\" (UID: \"323d9415-907b-4449-867a-2fb4659c9881\") " pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.669829 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5r4n\" (UniqueName: \"kubernetes.io/projected/323d9415-907b-4449-867a-2fb4659c9881-kube-api-access-m5r4n\") pod \"redhat-operators-rcqmp\" (UID: \"323d9415-907b-4449-867a-2fb4659c9881\") " pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.670085 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/323d9415-907b-4449-867a-2fb4659c9881-utilities\") pod \"redhat-operators-rcqmp\" (UID: \"323d9415-907b-4449-867a-2fb4659c9881\") " pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.720412 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zh6dj"] Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.723095 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.732209 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zh6dj"] Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.772955 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/323d9415-907b-4449-867a-2fb4659c9881-catalog-content\") pod \"redhat-operators-rcqmp\" (UID: \"323d9415-907b-4449-867a-2fb4659c9881\") " pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.773064 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5r4n\" (UniqueName: \"kubernetes.io/projected/323d9415-907b-4449-867a-2fb4659c9881-kube-api-access-m5r4n\") pod \"redhat-operators-rcqmp\" (UID: \"323d9415-907b-4449-867a-2fb4659c9881\") " pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.773156 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/323d9415-907b-4449-867a-2fb4659c9881-utilities\") pod \"redhat-operators-rcqmp\" (UID: \"323d9415-907b-4449-867a-2fb4659c9881\") " pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.776568 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/323d9415-907b-4449-867a-2fb4659c9881-utilities\") pod \"redhat-operators-rcqmp\" (UID: \"323d9415-907b-4449-867a-2fb4659c9881\") " pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.776716 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/323d9415-907b-4449-867a-2fb4659c9881-catalog-content\") pod \"redhat-operators-rcqmp\" (UID: \"323d9415-907b-4449-867a-2fb4659c9881\") " pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.803669 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5r4n\" (UniqueName: \"kubernetes.io/projected/323d9415-907b-4449-867a-2fb4659c9881-kube-api-access-m5r4n\") pod \"redhat-operators-rcqmp\" (UID: \"323d9415-907b-4449-867a-2fb4659c9881\") " pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.874727 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2qwl\" (UniqueName: \"kubernetes.io/projected/875aa6f7-331e-44b8-89f5-bacd67499a9d-kube-api-access-b2qwl\") pod \"community-operators-zh6dj\" (UID: \"875aa6f7-331e-44b8-89f5-bacd67499a9d\") " pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.874792 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/875aa6f7-331e-44b8-89f5-bacd67499a9d-catalog-content\") pod \"community-operators-zh6dj\" (UID: \"875aa6f7-331e-44b8-89f5-bacd67499a9d\") " pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.874820 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/875aa6f7-331e-44b8-89f5-bacd67499a9d-utilities\") pod \"community-operators-zh6dj\" (UID: \"875aa6f7-331e-44b8-89f5-bacd67499a9d\") " pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.905913 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.984729 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2qwl\" (UniqueName: \"kubernetes.io/projected/875aa6f7-331e-44b8-89f5-bacd67499a9d-kube-api-access-b2qwl\") pod \"community-operators-zh6dj\" (UID: \"875aa6f7-331e-44b8-89f5-bacd67499a9d\") " pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.984849 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/875aa6f7-331e-44b8-89f5-bacd67499a9d-catalog-content\") pod \"community-operators-zh6dj\" (UID: \"875aa6f7-331e-44b8-89f5-bacd67499a9d\") " pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.984880 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/875aa6f7-331e-44b8-89f5-bacd67499a9d-utilities\") pod \"community-operators-zh6dj\" (UID: \"875aa6f7-331e-44b8-89f5-bacd67499a9d\") " pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.985616 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/875aa6f7-331e-44b8-89f5-bacd67499a9d-utilities\") pod \"community-operators-zh6dj\" (UID: \"875aa6f7-331e-44b8-89f5-bacd67499a9d\") " pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:15:49 crc kubenswrapper[4771]: I1002 11:15:49.985758 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/875aa6f7-331e-44b8-89f5-bacd67499a9d-catalog-content\") pod \"community-operators-zh6dj\" (UID: \"875aa6f7-331e-44b8-89f5-bacd67499a9d\") " pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:15:50 crc kubenswrapper[4771]: I1002 11:15:50.003523 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2qwl\" (UniqueName: \"kubernetes.io/projected/875aa6f7-331e-44b8-89f5-bacd67499a9d-kube-api-access-b2qwl\") pod \"community-operators-zh6dj\" (UID: \"875aa6f7-331e-44b8-89f5-bacd67499a9d\") " pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:15:50 crc kubenswrapper[4771]: I1002 11:15:50.043648 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:15:50 crc kubenswrapper[4771]: I1002 11:15:50.535850 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rcqmp"] Oct 02 11:15:50 crc kubenswrapper[4771]: I1002 11:15:50.644338 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcqmp" event={"ID":"323d9415-907b-4449-867a-2fb4659c9881","Type":"ContainerStarted","Data":"224c006fa466a423ee32429d599fce4c4101390d02151c1883dbdea4bdf61509"} Oct 02 11:15:50 crc kubenswrapper[4771]: I1002 11:15:50.775507 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zh6dj"] Oct 02 11:15:50 crc kubenswrapper[4771]: W1002 11:15:50.803285 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod875aa6f7_331e_44b8_89f5_bacd67499a9d.slice/crio-883d52985ab7be735fffa5a77ff8b95cd5a844a75c1d8d42b837fe7f0c34ddd5 WatchSource:0}: Error finding container 883d52985ab7be735fffa5a77ff8b95cd5a844a75c1d8d42b837fe7f0c34ddd5: Status 404 returned error can't find the container with id 883d52985ab7be735fffa5a77ff8b95cd5a844a75c1d8d42b837fe7f0c34ddd5 Oct 02 11:15:51 crc kubenswrapper[4771]: I1002 11:15:51.655243 4771 generic.go:334] "Generic (PLEG): container finished" podID="323d9415-907b-4449-867a-2fb4659c9881" containerID="14545dd927e7f356411cb220cf6757ccc9048dcd4ef9c7a720409dc3a9773976" exitCode=0 Oct 02 11:15:51 crc kubenswrapper[4771]: I1002 11:15:51.655299 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcqmp" event={"ID":"323d9415-907b-4449-867a-2fb4659c9881","Type":"ContainerDied","Data":"14545dd927e7f356411cb220cf6757ccc9048dcd4ef9c7a720409dc3a9773976"} Oct 02 11:15:51 crc kubenswrapper[4771]: I1002 11:15:51.657845 4771 generic.go:334] "Generic (PLEG): container finished" podID="875aa6f7-331e-44b8-89f5-bacd67499a9d" containerID="4a68cabb5d814642a340fa03711d49be876d49df75de1999dc8c06ae4c7fd374" exitCode=0 Oct 02 11:15:51 crc kubenswrapper[4771]: I1002 11:15:51.657949 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zh6dj" event={"ID":"875aa6f7-331e-44b8-89f5-bacd67499a9d","Type":"ContainerDied","Data":"4a68cabb5d814642a340fa03711d49be876d49df75de1999dc8c06ae4c7fd374"} Oct 02 11:15:51 crc kubenswrapper[4771]: I1002 11:15:51.658037 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zh6dj" event={"ID":"875aa6f7-331e-44b8-89f5-bacd67499a9d","Type":"ContainerStarted","Data":"883d52985ab7be735fffa5a77ff8b95cd5a844a75c1d8d42b837fe7f0c34ddd5"} Oct 02 11:15:52 crc kubenswrapper[4771]: I1002 11:15:52.668778 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcqmp" event={"ID":"323d9415-907b-4449-867a-2fb4659c9881","Type":"ContainerStarted","Data":"f79c45d3937daaa83d82142f51f0859b943c16489a518346ae5aa6ec8e6693ff"} Oct 02 11:15:53 crc kubenswrapper[4771]: I1002 11:15:53.682304 4771 generic.go:334] "Generic (PLEG): container finished" podID="323d9415-907b-4449-867a-2fb4659c9881" containerID="f79c45d3937daaa83d82142f51f0859b943c16489a518346ae5aa6ec8e6693ff" exitCode=0 Oct 02 11:15:53 crc kubenswrapper[4771]: I1002 11:15:53.697839 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcqmp" event={"ID":"323d9415-907b-4449-867a-2fb4659c9881","Type":"ContainerDied","Data":"f79c45d3937daaa83d82142f51f0859b943c16489a518346ae5aa6ec8e6693ff"} Oct 02 11:15:53 crc kubenswrapper[4771]: I1002 11:15:53.697882 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zh6dj" event={"ID":"875aa6f7-331e-44b8-89f5-bacd67499a9d","Type":"ContainerStarted","Data":"083bfa286b7591b0ec5310f3d59e90a72605f5cf5d7cf2bdf389746e1c5facc8"} Oct 02 11:15:55 crc kubenswrapper[4771]: I1002 11:15:55.715856 4771 generic.go:334] "Generic (PLEG): container finished" podID="875aa6f7-331e-44b8-89f5-bacd67499a9d" containerID="083bfa286b7591b0ec5310f3d59e90a72605f5cf5d7cf2bdf389746e1c5facc8" exitCode=0 Oct 02 11:15:55 crc kubenswrapper[4771]: I1002 11:15:55.716341 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zh6dj" event={"ID":"875aa6f7-331e-44b8-89f5-bacd67499a9d","Type":"ContainerDied","Data":"083bfa286b7591b0ec5310f3d59e90a72605f5cf5d7cf2bdf389746e1c5facc8"} Oct 02 11:15:55 crc kubenswrapper[4771]: I1002 11:15:55.722107 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcqmp" event={"ID":"323d9415-907b-4449-867a-2fb4659c9881","Type":"ContainerStarted","Data":"2f4e52ad7e7e8ecd3189417452e05426bfa41f6d8dcacde23751e0324b9ce69a"} Oct 02 11:15:55 crc kubenswrapper[4771]: I1002 11:15:55.765676 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rcqmp" podStartSLOduration=3.360274419 podStartE2EDuration="6.765655135s" podCreationTimestamp="2025-10-02 11:15:49 +0000 UTC" firstStartedPulling="2025-10-02 11:15:51.657301552 +0000 UTC m=+5939.304986619" lastFinishedPulling="2025-10-02 11:15:55.062682268 +0000 UTC m=+5942.710367335" observedRunningTime="2025-10-02 11:15:55.757477933 +0000 UTC m=+5943.405163030" watchObservedRunningTime="2025-10-02 11:15:55.765655135 +0000 UTC m=+5943.413340212" Oct 02 11:15:56 crc kubenswrapper[4771]: I1002 11:15:56.734784 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zh6dj" event={"ID":"875aa6f7-331e-44b8-89f5-bacd67499a9d","Type":"ContainerStarted","Data":"2dec90ea6d428201ef8c2fb4a8eae9e3458b503db0ab2e9c89233fc1df529cc1"} Oct 02 11:15:56 crc kubenswrapper[4771]: I1002 11:15:56.759719 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zh6dj" podStartSLOduration=3.175188623 podStartE2EDuration="7.759693954s" podCreationTimestamp="2025-10-02 11:15:49 +0000 UTC" firstStartedPulling="2025-10-02 11:15:51.661710796 +0000 UTC m=+5939.309395863" lastFinishedPulling="2025-10-02 11:15:56.246216127 +0000 UTC m=+5943.893901194" observedRunningTime="2025-10-02 11:15:56.752575749 +0000 UTC m=+5944.400260816" watchObservedRunningTime="2025-10-02 11:15:56.759693954 +0000 UTC m=+5944.407379021" Oct 02 11:15:58 crc kubenswrapper[4771]: I1002 11:15:58.683653 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:15:58 crc kubenswrapper[4771]: E1002 11:15:58.685001 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:15:59 crc kubenswrapper[4771]: I1002 11:15:59.907024 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:15:59 crc kubenswrapper[4771]: I1002 11:15:59.907381 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:16:00 crc kubenswrapper[4771]: I1002 11:16:00.044560 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:16:00 crc kubenswrapper[4771]: I1002 11:16:00.045028 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:16:00 crc kubenswrapper[4771]: I1002 11:16:00.102455 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:16:00 crc kubenswrapper[4771]: I1002 11:16:00.959177 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rcqmp" podUID="323d9415-907b-4449-867a-2fb4659c9881" containerName="registry-server" probeResult="failure" output=< Oct 02 11:16:00 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:16:00 crc kubenswrapper[4771]: > Oct 02 11:16:01 crc kubenswrapper[4771]: I1002 11:16:01.851465 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:16:02 crc kubenswrapper[4771]: I1002 11:16:02.308004 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zh6dj"] Oct 02 11:16:03 crc kubenswrapper[4771]: I1002 11:16:03.836432 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zh6dj" podUID="875aa6f7-331e-44b8-89f5-bacd67499a9d" containerName="registry-server" containerID="cri-o://2dec90ea6d428201ef8c2fb4a8eae9e3458b503db0ab2e9c89233fc1df529cc1" gracePeriod=2 Oct 02 11:16:04 crc kubenswrapper[4771]: I1002 11:16:04.862613 4771 generic.go:334] "Generic (PLEG): container finished" podID="875aa6f7-331e-44b8-89f5-bacd67499a9d" containerID="2dec90ea6d428201ef8c2fb4a8eae9e3458b503db0ab2e9c89233fc1df529cc1" exitCode=0 Oct 02 11:16:04 crc kubenswrapper[4771]: I1002 11:16:04.862715 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zh6dj" event={"ID":"875aa6f7-331e-44b8-89f5-bacd67499a9d","Type":"ContainerDied","Data":"2dec90ea6d428201ef8c2fb4a8eae9e3458b503db0ab2e9c89233fc1df529cc1"} Oct 02 11:16:04 crc kubenswrapper[4771]: I1002 11:16:04.863078 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zh6dj" event={"ID":"875aa6f7-331e-44b8-89f5-bacd67499a9d","Type":"ContainerDied","Data":"883d52985ab7be735fffa5a77ff8b95cd5a844a75c1d8d42b837fe7f0c34ddd5"} Oct 02 11:16:04 crc kubenswrapper[4771]: I1002 11:16:04.863095 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="883d52985ab7be735fffa5a77ff8b95cd5a844a75c1d8d42b837fe7f0c34ddd5" Oct 02 11:16:04 crc kubenswrapper[4771]: I1002 11:16:04.913447 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:16:05 crc kubenswrapper[4771]: I1002 11:16:05.006614 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/875aa6f7-331e-44b8-89f5-bacd67499a9d-catalog-content\") pod \"875aa6f7-331e-44b8-89f5-bacd67499a9d\" (UID: \"875aa6f7-331e-44b8-89f5-bacd67499a9d\") " Oct 02 11:16:05 crc kubenswrapper[4771]: I1002 11:16:05.006818 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/875aa6f7-331e-44b8-89f5-bacd67499a9d-utilities\") pod \"875aa6f7-331e-44b8-89f5-bacd67499a9d\" (UID: \"875aa6f7-331e-44b8-89f5-bacd67499a9d\") " Oct 02 11:16:05 crc kubenswrapper[4771]: I1002 11:16:05.006885 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2qwl\" (UniqueName: \"kubernetes.io/projected/875aa6f7-331e-44b8-89f5-bacd67499a9d-kube-api-access-b2qwl\") pod \"875aa6f7-331e-44b8-89f5-bacd67499a9d\" (UID: \"875aa6f7-331e-44b8-89f5-bacd67499a9d\") " Oct 02 11:16:05 crc kubenswrapper[4771]: I1002 11:16:05.009660 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/875aa6f7-331e-44b8-89f5-bacd67499a9d-utilities" (OuterVolumeSpecName: "utilities") pod "875aa6f7-331e-44b8-89f5-bacd67499a9d" (UID: "875aa6f7-331e-44b8-89f5-bacd67499a9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:16:05 crc kubenswrapper[4771]: I1002 11:16:05.014338 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/875aa6f7-331e-44b8-89f5-bacd67499a9d-kube-api-access-b2qwl" (OuterVolumeSpecName: "kube-api-access-b2qwl") pod "875aa6f7-331e-44b8-89f5-bacd67499a9d" (UID: "875aa6f7-331e-44b8-89f5-bacd67499a9d"). InnerVolumeSpecName "kube-api-access-b2qwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:05 crc kubenswrapper[4771]: I1002 11:16:05.062649 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/875aa6f7-331e-44b8-89f5-bacd67499a9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "875aa6f7-331e-44b8-89f5-bacd67499a9d" (UID: "875aa6f7-331e-44b8-89f5-bacd67499a9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:16:05 crc kubenswrapper[4771]: I1002 11:16:05.110052 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/875aa6f7-331e-44b8-89f5-bacd67499a9d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:05 crc kubenswrapper[4771]: I1002 11:16:05.110105 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/875aa6f7-331e-44b8-89f5-bacd67499a9d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:05 crc kubenswrapper[4771]: I1002 11:16:05.110121 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2qwl\" (UniqueName: \"kubernetes.io/projected/875aa6f7-331e-44b8-89f5-bacd67499a9d-kube-api-access-b2qwl\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:05 crc kubenswrapper[4771]: I1002 11:16:05.875109 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zh6dj" Oct 02 11:16:05 crc kubenswrapper[4771]: I1002 11:16:05.906090 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zh6dj"] Oct 02 11:16:05 crc kubenswrapper[4771]: I1002 11:16:05.918479 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zh6dj"] Oct 02 11:16:07 crc kubenswrapper[4771]: I1002 11:16:07.697666 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="875aa6f7-331e-44b8-89f5-bacd67499a9d" path="/var/lib/kubelet/pods/875aa6f7-331e-44b8-89f5-bacd67499a9d/volumes" Oct 02 11:16:10 crc kubenswrapper[4771]: I1002 11:16:10.963492 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rcqmp" podUID="323d9415-907b-4449-867a-2fb4659c9881" containerName="registry-server" probeResult="failure" output=< Oct 02 11:16:10 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:16:10 crc kubenswrapper[4771]: > Oct 02 11:16:12 crc kubenswrapper[4771]: I1002 11:16:12.681210 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:16:12 crc kubenswrapper[4771]: E1002 11:16:12.682066 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:16:19 crc kubenswrapper[4771]: I1002 11:16:19.967418 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:16:20 crc kubenswrapper[4771]: I1002 11:16:20.023055 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:16:20 crc kubenswrapper[4771]: I1002 11:16:20.730076 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rcqmp"] Oct 02 11:16:21 crc kubenswrapper[4771]: I1002 11:16:21.052900 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rcqmp" podUID="323d9415-907b-4449-867a-2fb4659c9881" containerName="registry-server" containerID="cri-o://2f4e52ad7e7e8ecd3189417452e05426bfa41f6d8dcacde23751e0324b9ce69a" gracePeriod=2 Oct 02 11:16:21 crc kubenswrapper[4771]: I1002 11:16:21.623721 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:16:21 crc kubenswrapper[4771]: I1002 11:16:21.706114 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5r4n\" (UniqueName: \"kubernetes.io/projected/323d9415-907b-4449-867a-2fb4659c9881-kube-api-access-m5r4n\") pod \"323d9415-907b-4449-867a-2fb4659c9881\" (UID: \"323d9415-907b-4449-867a-2fb4659c9881\") " Oct 02 11:16:21 crc kubenswrapper[4771]: I1002 11:16:21.706395 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/323d9415-907b-4449-867a-2fb4659c9881-utilities\") pod \"323d9415-907b-4449-867a-2fb4659c9881\" (UID: \"323d9415-907b-4449-867a-2fb4659c9881\") " Oct 02 11:16:21 crc kubenswrapper[4771]: I1002 11:16:21.706468 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/323d9415-907b-4449-867a-2fb4659c9881-catalog-content\") pod \"323d9415-907b-4449-867a-2fb4659c9881\" (UID: \"323d9415-907b-4449-867a-2fb4659c9881\") " Oct 02 11:16:21 crc kubenswrapper[4771]: I1002 11:16:21.710299 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/323d9415-907b-4449-867a-2fb4659c9881-utilities" (OuterVolumeSpecName: "utilities") pod "323d9415-907b-4449-867a-2fb4659c9881" (UID: "323d9415-907b-4449-867a-2fb4659c9881"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:16:21 crc kubenswrapper[4771]: I1002 11:16:21.718422 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/323d9415-907b-4449-867a-2fb4659c9881-kube-api-access-m5r4n" (OuterVolumeSpecName: "kube-api-access-m5r4n") pod "323d9415-907b-4449-867a-2fb4659c9881" (UID: "323d9415-907b-4449-867a-2fb4659c9881"). InnerVolumeSpecName "kube-api-access-m5r4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:21 crc kubenswrapper[4771]: I1002 11:16:21.790349 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/323d9415-907b-4449-867a-2fb4659c9881-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "323d9415-907b-4449-867a-2fb4659c9881" (UID: "323d9415-907b-4449-867a-2fb4659c9881"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:16:21 crc kubenswrapper[4771]: I1002 11:16:21.809259 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5r4n\" (UniqueName: \"kubernetes.io/projected/323d9415-907b-4449-867a-2fb4659c9881-kube-api-access-m5r4n\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:21 crc kubenswrapper[4771]: I1002 11:16:21.809477 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/323d9415-907b-4449-867a-2fb4659c9881-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:21 crc kubenswrapper[4771]: I1002 11:16:21.809579 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/323d9415-907b-4449-867a-2fb4659c9881-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.064035 4771 generic.go:334] "Generic (PLEG): container finished" podID="323d9415-907b-4449-867a-2fb4659c9881" containerID="2f4e52ad7e7e8ecd3189417452e05426bfa41f6d8dcacde23751e0324b9ce69a" exitCode=0 Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.064085 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcqmp" Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.064105 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcqmp" event={"ID":"323d9415-907b-4449-867a-2fb4659c9881","Type":"ContainerDied","Data":"2f4e52ad7e7e8ecd3189417452e05426bfa41f6d8dcacde23751e0324b9ce69a"} Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.065182 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcqmp" event={"ID":"323d9415-907b-4449-867a-2fb4659c9881","Type":"ContainerDied","Data":"224c006fa466a423ee32429d599fce4c4101390d02151c1883dbdea4bdf61509"} Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.065211 4771 scope.go:117] "RemoveContainer" containerID="2f4e52ad7e7e8ecd3189417452e05426bfa41f6d8dcacde23751e0324b9ce69a" Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.100006 4771 scope.go:117] "RemoveContainer" containerID="f79c45d3937daaa83d82142f51f0859b943c16489a518346ae5aa6ec8e6693ff" Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.102495 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rcqmp"] Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.118453 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rcqmp"] Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.125847 4771 scope.go:117] "RemoveContainer" containerID="14545dd927e7f356411cb220cf6757ccc9048dcd4ef9c7a720409dc3a9773976" Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.179737 4771 scope.go:117] "RemoveContainer" containerID="2f4e52ad7e7e8ecd3189417452e05426bfa41f6d8dcacde23751e0324b9ce69a" Oct 02 11:16:22 crc kubenswrapper[4771]: E1002 11:16:22.180600 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f4e52ad7e7e8ecd3189417452e05426bfa41f6d8dcacde23751e0324b9ce69a\": container with ID starting with 2f4e52ad7e7e8ecd3189417452e05426bfa41f6d8dcacde23751e0324b9ce69a not found: ID does not exist" containerID="2f4e52ad7e7e8ecd3189417452e05426bfa41f6d8dcacde23751e0324b9ce69a" Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.180643 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f4e52ad7e7e8ecd3189417452e05426bfa41f6d8dcacde23751e0324b9ce69a"} err="failed to get container status \"2f4e52ad7e7e8ecd3189417452e05426bfa41f6d8dcacde23751e0324b9ce69a\": rpc error: code = NotFound desc = could not find container \"2f4e52ad7e7e8ecd3189417452e05426bfa41f6d8dcacde23751e0324b9ce69a\": container with ID starting with 2f4e52ad7e7e8ecd3189417452e05426bfa41f6d8dcacde23751e0324b9ce69a not found: ID does not exist" Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.180671 4771 scope.go:117] "RemoveContainer" containerID="f79c45d3937daaa83d82142f51f0859b943c16489a518346ae5aa6ec8e6693ff" Oct 02 11:16:22 crc kubenswrapper[4771]: E1002 11:16:22.181210 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f79c45d3937daaa83d82142f51f0859b943c16489a518346ae5aa6ec8e6693ff\": container with ID starting with f79c45d3937daaa83d82142f51f0859b943c16489a518346ae5aa6ec8e6693ff not found: ID does not exist" containerID="f79c45d3937daaa83d82142f51f0859b943c16489a518346ae5aa6ec8e6693ff" Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.181235 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f79c45d3937daaa83d82142f51f0859b943c16489a518346ae5aa6ec8e6693ff"} err="failed to get container status \"f79c45d3937daaa83d82142f51f0859b943c16489a518346ae5aa6ec8e6693ff\": rpc error: code = NotFound desc = could not find container \"f79c45d3937daaa83d82142f51f0859b943c16489a518346ae5aa6ec8e6693ff\": container with ID starting with f79c45d3937daaa83d82142f51f0859b943c16489a518346ae5aa6ec8e6693ff not found: ID does not exist" Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.181253 4771 scope.go:117] "RemoveContainer" containerID="14545dd927e7f356411cb220cf6757ccc9048dcd4ef9c7a720409dc3a9773976" Oct 02 11:16:22 crc kubenswrapper[4771]: E1002 11:16:22.181721 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14545dd927e7f356411cb220cf6757ccc9048dcd4ef9c7a720409dc3a9773976\": container with ID starting with 14545dd927e7f356411cb220cf6757ccc9048dcd4ef9c7a720409dc3a9773976 not found: ID does not exist" containerID="14545dd927e7f356411cb220cf6757ccc9048dcd4ef9c7a720409dc3a9773976" Oct 02 11:16:22 crc kubenswrapper[4771]: I1002 11:16:22.181750 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14545dd927e7f356411cb220cf6757ccc9048dcd4ef9c7a720409dc3a9773976"} err="failed to get container status \"14545dd927e7f356411cb220cf6757ccc9048dcd4ef9c7a720409dc3a9773976\": rpc error: code = NotFound desc = could not find container \"14545dd927e7f356411cb220cf6757ccc9048dcd4ef9c7a720409dc3a9773976\": container with ID starting with 14545dd927e7f356411cb220cf6757ccc9048dcd4ef9c7a720409dc3a9773976 not found: ID does not exist" Oct 02 11:16:23 crc kubenswrapper[4771]: I1002 11:16:23.699023 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="323d9415-907b-4449-867a-2fb4659c9881" path="/var/lib/kubelet/pods/323d9415-907b-4449-867a-2fb4659c9881/volumes" Oct 02 11:16:27 crc kubenswrapper[4771]: I1002 11:16:27.681636 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:16:27 crc kubenswrapper[4771]: E1002 11:16:27.682608 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:16:39 crc kubenswrapper[4771]: I1002 11:16:39.681519 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:16:39 crc kubenswrapper[4771]: E1002 11:16:39.682529 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:16:53 crc kubenswrapper[4771]: I1002 11:16:53.691933 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:16:53 crc kubenswrapper[4771]: E1002 11:16:53.692645 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:17:04 crc kubenswrapper[4771]: I1002 11:17:04.681606 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:17:04 crc kubenswrapper[4771]: E1002 11:17:04.682703 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:17:17 crc kubenswrapper[4771]: I1002 11:17:17.681645 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:17:18 crc kubenswrapper[4771]: I1002 11:17:18.710327 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"5f3b64f0ba9c14fd906fe67d63409842d8b833922da63c8e0956c0158ecdb611"} Oct 02 11:19:42 crc kubenswrapper[4771]: I1002 11:19:42.146375 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:19:42 crc kubenswrapper[4771]: I1002 11:19:42.147414 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:20:12 crc kubenswrapper[4771]: I1002 11:20:12.146072 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:20:12 crc kubenswrapper[4771]: I1002 11:20:12.149056 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:20:42 crc kubenswrapper[4771]: I1002 11:20:42.145862 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:20:42 crc kubenswrapper[4771]: I1002 11:20:42.146360 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:20:42 crc kubenswrapper[4771]: I1002 11:20:42.146403 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 11:20:42 crc kubenswrapper[4771]: I1002 11:20:42.147281 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5f3b64f0ba9c14fd906fe67d63409842d8b833922da63c8e0956c0158ecdb611"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:20:42 crc kubenswrapper[4771]: I1002 11:20:42.147325 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://5f3b64f0ba9c14fd906fe67d63409842d8b833922da63c8e0956c0158ecdb611" gracePeriod=600 Oct 02 11:20:43 crc kubenswrapper[4771]: I1002 11:20:43.120615 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="5f3b64f0ba9c14fd906fe67d63409842d8b833922da63c8e0956c0158ecdb611" exitCode=0 Oct 02 11:20:43 crc kubenswrapper[4771]: I1002 11:20:43.120683 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"5f3b64f0ba9c14fd906fe67d63409842d8b833922da63c8e0956c0158ecdb611"} Oct 02 11:20:43 crc kubenswrapper[4771]: I1002 11:20:43.121199 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc"} Oct 02 11:20:43 crc kubenswrapper[4771]: I1002 11:20:43.121237 4771 scope.go:117] "RemoveContainer" containerID="b8470898aba4f2679f6781553b31efba52128424ffc6797d6883de477ac1a21a" Oct 02 11:21:32 crc kubenswrapper[4771]: I1002 11:21:32.867033 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nn4cs"] Oct 02 11:21:32 crc kubenswrapper[4771]: E1002 11:21:32.868412 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875aa6f7-331e-44b8-89f5-bacd67499a9d" containerName="extract-utilities" Oct 02 11:21:32 crc kubenswrapper[4771]: I1002 11:21:32.868433 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="875aa6f7-331e-44b8-89f5-bacd67499a9d" containerName="extract-utilities" Oct 02 11:21:32 crc kubenswrapper[4771]: E1002 11:21:32.868470 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875aa6f7-331e-44b8-89f5-bacd67499a9d" containerName="registry-server" Oct 02 11:21:32 crc kubenswrapper[4771]: I1002 11:21:32.868479 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="875aa6f7-331e-44b8-89f5-bacd67499a9d" containerName="registry-server" Oct 02 11:21:32 crc kubenswrapper[4771]: E1002 11:21:32.868496 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="323d9415-907b-4449-867a-2fb4659c9881" containerName="registry-server" Oct 02 11:21:32 crc kubenswrapper[4771]: I1002 11:21:32.868503 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="323d9415-907b-4449-867a-2fb4659c9881" containerName="registry-server" Oct 02 11:21:32 crc kubenswrapper[4771]: E1002 11:21:32.868523 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="323d9415-907b-4449-867a-2fb4659c9881" containerName="extract-content" Oct 02 11:21:32 crc kubenswrapper[4771]: I1002 11:21:32.868532 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="323d9415-907b-4449-867a-2fb4659c9881" containerName="extract-content" Oct 02 11:21:32 crc kubenswrapper[4771]: E1002 11:21:32.868551 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="323d9415-907b-4449-867a-2fb4659c9881" containerName="extract-utilities" Oct 02 11:21:32 crc kubenswrapper[4771]: I1002 11:21:32.868559 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="323d9415-907b-4449-867a-2fb4659c9881" containerName="extract-utilities" Oct 02 11:21:32 crc kubenswrapper[4771]: E1002 11:21:32.868572 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875aa6f7-331e-44b8-89f5-bacd67499a9d" containerName="extract-content" Oct 02 11:21:32 crc kubenswrapper[4771]: I1002 11:21:32.868582 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="875aa6f7-331e-44b8-89f5-bacd67499a9d" containerName="extract-content" Oct 02 11:21:32 crc kubenswrapper[4771]: I1002 11:21:32.868855 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="875aa6f7-331e-44b8-89f5-bacd67499a9d" containerName="registry-server" Oct 02 11:21:32 crc kubenswrapper[4771]: I1002 11:21:32.868880 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="323d9415-907b-4449-867a-2fb4659c9881" containerName="registry-server" Oct 02 11:21:32 crc kubenswrapper[4771]: I1002 11:21:32.870930 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:32 crc kubenswrapper[4771]: I1002 11:21:32.885852 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nn4cs"] Oct 02 11:21:33 crc kubenswrapper[4771]: I1002 11:21:33.064200 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbn2r\" (UniqueName: \"kubernetes.io/projected/9f334abc-1b22-4577-821b-ae4d10f2a1b6-kube-api-access-gbn2r\") pod \"certified-operators-nn4cs\" (UID: \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\") " pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:33 crc kubenswrapper[4771]: I1002 11:21:33.064807 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f334abc-1b22-4577-821b-ae4d10f2a1b6-utilities\") pod \"certified-operators-nn4cs\" (UID: \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\") " pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:33 crc kubenswrapper[4771]: I1002 11:21:33.065025 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f334abc-1b22-4577-821b-ae4d10f2a1b6-catalog-content\") pod \"certified-operators-nn4cs\" (UID: \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\") " pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:33 crc kubenswrapper[4771]: I1002 11:21:33.167492 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f334abc-1b22-4577-821b-ae4d10f2a1b6-catalog-content\") pod \"certified-operators-nn4cs\" (UID: \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\") " pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:33 crc kubenswrapper[4771]: I1002 11:21:33.167997 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f334abc-1b22-4577-821b-ae4d10f2a1b6-catalog-content\") pod \"certified-operators-nn4cs\" (UID: \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\") " pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:33 crc kubenswrapper[4771]: I1002 11:21:33.168006 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbn2r\" (UniqueName: \"kubernetes.io/projected/9f334abc-1b22-4577-821b-ae4d10f2a1b6-kube-api-access-gbn2r\") pod \"certified-operators-nn4cs\" (UID: \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\") " pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:33 crc kubenswrapper[4771]: I1002 11:21:33.168199 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f334abc-1b22-4577-821b-ae4d10f2a1b6-utilities\") pod \"certified-operators-nn4cs\" (UID: \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\") " pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:33 crc kubenswrapper[4771]: I1002 11:21:33.168567 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f334abc-1b22-4577-821b-ae4d10f2a1b6-utilities\") pod \"certified-operators-nn4cs\" (UID: \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\") " pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:33 crc kubenswrapper[4771]: I1002 11:21:33.208091 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbn2r\" (UniqueName: \"kubernetes.io/projected/9f334abc-1b22-4577-821b-ae4d10f2a1b6-kube-api-access-gbn2r\") pod \"certified-operators-nn4cs\" (UID: \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\") " pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:33 crc kubenswrapper[4771]: I1002 11:21:33.499642 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:34 crc kubenswrapper[4771]: I1002 11:21:34.106967 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nn4cs"] Oct 02 11:21:34 crc kubenswrapper[4771]: I1002 11:21:34.684397 4771 generic.go:334] "Generic (PLEG): container finished" podID="9f334abc-1b22-4577-821b-ae4d10f2a1b6" containerID="504e3f318dc174a79a40e5ef5582594b7afdaf92ff4dfce6e58d691bddadf477" exitCode=0 Oct 02 11:21:34 crc kubenswrapper[4771]: I1002 11:21:34.684726 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nn4cs" event={"ID":"9f334abc-1b22-4577-821b-ae4d10f2a1b6","Type":"ContainerDied","Data":"504e3f318dc174a79a40e5ef5582594b7afdaf92ff4dfce6e58d691bddadf477"} Oct 02 11:21:34 crc kubenswrapper[4771]: I1002 11:21:34.684849 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nn4cs" event={"ID":"9f334abc-1b22-4577-821b-ae4d10f2a1b6","Type":"ContainerStarted","Data":"5a7d6ad87af3dabc818d16212a827360077a06672d6d7096f918d814a485798d"} Oct 02 11:21:34 crc kubenswrapper[4771]: I1002 11:21:34.686578 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:21:39 crc kubenswrapper[4771]: I1002 11:21:39.777423 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nn4cs" event={"ID":"9f334abc-1b22-4577-821b-ae4d10f2a1b6","Type":"ContainerStarted","Data":"52306ddaca8cec6a4b56a8ab4d52dd822c0425dd55554248fb264ddf6dcf06a0"} Oct 02 11:21:43 crc kubenswrapper[4771]: I1002 11:21:43.818633 4771 generic.go:334] "Generic (PLEG): container finished" podID="9f334abc-1b22-4577-821b-ae4d10f2a1b6" containerID="52306ddaca8cec6a4b56a8ab4d52dd822c0425dd55554248fb264ddf6dcf06a0" exitCode=0 Oct 02 11:21:43 crc kubenswrapper[4771]: I1002 11:21:43.818727 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nn4cs" event={"ID":"9f334abc-1b22-4577-821b-ae4d10f2a1b6","Type":"ContainerDied","Data":"52306ddaca8cec6a4b56a8ab4d52dd822c0425dd55554248fb264ddf6dcf06a0"} Oct 02 11:21:46 crc kubenswrapper[4771]: I1002 11:21:46.855030 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nn4cs" event={"ID":"9f334abc-1b22-4577-821b-ae4d10f2a1b6","Type":"ContainerStarted","Data":"252b28bc565c7e308fb9d0518a47ffb09d74148b5b00345ab4e580527d76b8d3"} Oct 02 11:21:46 crc kubenswrapper[4771]: I1002 11:21:46.883302 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nn4cs" podStartSLOduration=3.790436343 podStartE2EDuration="14.883281351s" podCreationTimestamp="2025-10-02 11:21:32 +0000 UTC" firstStartedPulling="2025-10-02 11:21:34.686357753 +0000 UTC m=+6282.334042820" lastFinishedPulling="2025-10-02 11:21:45.779202761 +0000 UTC m=+6293.426887828" observedRunningTime="2025-10-02 11:21:46.873358504 +0000 UTC m=+6294.521043571" watchObservedRunningTime="2025-10-02 11:21:46.883281351 +0000 UTC m=+6294.530966418" Oct 02 11:21:53 crc kubenswrapper[4771]: I1002 11:21:53.501598 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:53 crc kubenswrapper[4771]: I1002 11:21:53.502214 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:53 crc kubenswrapper[4771]: I1002 11:21:53.557982 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:53 crc kubenswrapper[4771]: I1002 11:21:53.975360 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:54 crc kubenswrapper[4771]: I1002 11:21:54.057642 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nn4cs"] Oct 02 11:21:55 crc kubenswrapper[4771]: I1002 11:21:55.944934 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nn4cs" podUID="9f334abc-1b22-4577-821b-ae4d10f2a1b6" containerName="registry-server" containerID="cri-o://252b28bc565c7e308fb9d0518a47ffb09d74148b5b00345ab4e580527d76b8d3" gracePeriod=2 Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.467293 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.504687 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f334abc-1b22-4577-821b-ae4d10f2a1b6-catalog-content\") pod \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\" (UID: \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\") " Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.505041 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbn2r\" (UniqueName: \"kubernetes.io/projected/9f334abc-1b22-4577-821b-ae4d10f2a1b6-kube-api-access-gbn2r\") pod \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\" (UID: \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\") " Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.505115 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f334abc-1b22-4577-821b-ae4d10f2a1b6-utilities\") pod \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\" (UID: \"9f334abc-1b22-4577-821b-ae4d10f2a1b6\") " Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.506336 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f334abc-1b22-4577-821b-ae4d10f2a1b6-utilities" (OuterVolumeSpecName: "utilities") pod "9f334abc-1b22-4577-821b-ae4d10f2a1b6" (UID: "9f334abc-1b22-4577-821b-ae4d10f2a1b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.520416 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f334abc-1b22-4577-821b-ae4d10f2a1b6-kube-api-access-gbn2r" (OuterVolumeSpecName: "kube-api-access-gbn2r") pod "9f334abc-1b22-4577-821b-ae4d10f2a1b6" (UID: "9f334abc-1b22-4577-821b-ae4d10f2a1b6"). InnerVolumeSpecName "kube-api-access-gbn2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.551344 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f334abc-1b22-4577-821b-ae4d10f2a1b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f334abc-1b22-4577-821b-ae4d10f2a1b6" (UID: "9f334abc-1b22-4577-821b-ae4d10f2a1b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.607577 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbn2r\" (UniqueName: \"kubernetes.io/projected/9f334abc-1b22-4577-821b-ae4d10f2a1b6-kube-api-access-gbn2r\") on node \"crc\" DevicePath \"\"" Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.607611 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f334abc-1b22-4577-821b-ae4d10f2a1b6-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.607621 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f334abc-1b22-4577-821b-ae4d10f2a1b6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.956968 4771 generic.go:334] "Generic (PLEG): container finished" podID="9f334abc-1b22-4577-821b-ae4d10f2a1b6" containerID="252b28bc565c7e308fb9d0518a47ffb09d74148b5b00345ab4e580527d76b8d3" exitCode=0 Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.957015 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nn4cs" event={"ID":"9f334abc-1b22-4577-821b-ae4d10f2a1b6","Type":"ContainerDied","Data":"252b28bc565c7e308fb9d0518a47ffb09d74148b5b00345ab4e580527d76b8d3"} Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.957046 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nn4cs" event={"ID":"9f334abc-1b22-4577-821b-ae4d10f2a1b6","Type":"ContainerDied","Data":"5a7d6ad87af3dabc818d16212a827360077a06672d6d7096f918d814a485798d"} Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.957066 4771 scope.go:117] "RemoveContainer" containerID="252b28bc565c7e308fb9d0518a47ffb09d74148b5b00345ab4e580527d76b8d3" Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.957238 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nn4cs" Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.987505 4771 scope.go:117] "RemoveContainer" containerID="52306ddaca8cec6a4b56a8ab4d52dd822c0425dd55554248fb264ddf6dcf06a0" Oct 02 11:21:56 crc kubenswrapper[4771]: I1002 11:21:56.993222 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nn4cs"] Oct 02 11:21:57 crc kubenswrapper[4771]: I1002 11:21:57.006882 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nn4cs"] Oct 02 11:21:57 crc kubenswrapper[4771]: I1002 11:21:57.019162 4771 scope.go:117] "RemoveContainer" containerID="504e3f318dc174a79a40e5ef5582594b7afdaf92ff4dfce6e58d691bddadf477" Oct 02 11:21:57 crc kubenswrapper[4771]: I1002 11:21:57.071955 4771 scope.go:117] "RemoveContainer" containerID="252b28bc565c7e308fb9d0518a47ffb09d74148b5b00345ab4e580527d76b8d3" Oct 02 11:21:57 crc kubenswrapper[4771]: E1002 11:21:57.072419 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"252b28bc565c7e308fb9d0518a47ffb09d74148b5b00345ab4e580527d76b8d3\": container with ID starting with 252b28bc565c7e308fb9d0518a47ffb09d74148b5b00345ab4e580527d76b8d3 not found: ID does not exist" containerID="252b28bc565c7e308fb9d0518a47ffb09d74148b5b00345ab4e580527d76b8d3" Oct 02 11:21:57 crc kubenswrapper[4771]: I1002 11:21:57.072458 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"252b28bc565c7e308fb9d0518a47ffb09d74148b5b00345ab4e580527d76b8d3"} err="failed to get container status \"252b28bc565c7e308fb9d0518a47ffb09d74148b5b00345ab4e580527d76b8d3\": rpc error: code = NotFound desc = could not find container \"252b28bc565c7e308fb9d0518a47ffb09d74148b5b00345ab4e580527d76b8d3\": container with ID starting with 252b28bc565c7e308fb9d0518a47ffb09d74148b5b00345ab4e580527d76b8d3 not found: ID does not exist" Oct 02 11:21:57 crc kubenswrapper[4771]: I1002 11:21:57.072484 4771 scope.go:117] "RemoveContainer" containerID="52306ddaca8cec6a4b56a8ab4d52dd822c0425dd55554248fb264ddf6dcf06a0" Oct 02 11:21:57 crc kubenswrapper[4771]: E1002 11:21:57.073172 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52306ddaca8cec6a4b56a8ab4d52dd822c0425dd55554248fb264ddf6dcf06a0\": container with ID starting with 52306ddaca8cec6a4b56a8ab4d52dd822c0425dd55554248fb264ddf6dcf06a0 not found: ID does not exist" containerID="52306ddaca8cec6a4b56a8ab4d52dd822c0425dd55554248fb264ddf6dcf06a0" Oct 02 11:21:57 crc kubenswrapper[4771]: I1002 11:21:57.073217 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52306ddaca8cec6a4b56a8ab4d52dd822c0425dd55554248fb264ddf6dcf06a0"} err="failed to get container status \"52306ddaca8cec6a4b56a8ab4d52dd822c0425dd55554248fb264ddf6dcf06a0\": rpc error: code = NotFound desc = could not find container \"52306ddaca8cec6a4b56a8ab4d52dd822c0425dd55554248fb264ddf6dcf06a0\": container with ID starting with 52306ddaca8cec6a4b56a8ab4d52dd822c0425dd55554248fb264ddf6dcf06a0 not found: ID does not exist" Oct 02 11:21:57 crc kubenswrapper[4771]: I1002 11:21:57.073247 4771 scope.go:117] "RemoveContainer" containerID="504e3f318dc174a79a40e5ef5582594b7afdaf92ff4dfce6e58d691bddadf477" Oct 02 11:21:57 crc kubenswrapper[4771]: E1002 11:21:57.073479 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"504e3f318dc174a79a40e5ef5582594b7afdaf92ff4dfce6e58d691bddadf477\": container with ID starting with 504e3f318dc174a79a40e5ef5582594b7afdaf92ff4dfce6e58d691bddadf477 not found: ID does not exist" containerID="504e3f318dc174a79a40e5ef5582594b7afdaf92ff4dfce6e58d691bddadf477" Oct 02 11:21:57 crc kubenswrapper[4771]: I1002 11:21:57.073513 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"504e3f318dc174a79a40e5ef5582594b7afdaf92ff4dfce6e58d691bddadf477"} err="failed to get container status \"504e3f318dc174a79a40e5ef5582594b7afdaf92ff4dfce6e58d691bddadf477\": rpc error: code = NotFound desc = could not find container \"504e3f318dc174a79a40e5ef5582594b7afdaf92ff4dfce6e58d691bddadf477\": container with ID starting with 504e3f318dc174a79a40e5ef5582594b7afdaf92ff4dfce6e58d691bddadf477 not found: ID does not exist" Oct 02 11:21:57 crc kubenswrapper[4771]: I1002 11:21:57.699515 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f334abc-1b22-4577-821b-ae4d10f2a1b6" path="/var/lib/kubelet/pods/9f334abc-1b22-4577-821b-ae4d10f2a1b6/volumes" Oct 02 11:22:06 crc kubenswrapper[4771]: I1002 11:22:06.789303 4771 scope.go:117] "RemoveContainer" containerID="083bfa286b7591b0ec5310f3d59e90a72605f5cf5d7cf2bdf389746e1c5facc8" Oct 02 11:22:06 crc kubenswrapper[4771]: I1002 11:22:06.830003 4771 scope.go:117] "RemoveContainer" containerID="2dec90ea6d428201ef8c2fb4a8eae9e3458b503db0ab2e9c89233fc1df529cc1" Oct 02 11:22:06 crc kubenswrapper[4771]: I1002 11:22:06.884075 4771 scope.go:117] "RemoveContainer" containerID="4a68cabb5d814642a340fa03711d49be876d49df75de1999dc8c06ae4c7fd374" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.407756 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-482nv"] Oct 02 11:22:36 crc kubenswrapper[4771]: E1002 11:22:36.411288 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f334abc-1b22-4577-821b-ae4d10f2a1b6" containerName="registry-server" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.411307 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f334abc-1b22-4577-821b-ae4d10f2a1b6" containerName="registry-server" Oct 02 11:22:36 crc kubenswrapper[4771]: E1002 11:22:36.411324 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f334abc-1b22-4577-821b-ae4d10f2a1b6" containerName="extract-utilities" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.411331 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f334abc-1b22-4577-821b-ae4d10f2a1b6" containerName="extract-utilities" Oct 02 11:22:36 crc kubenswrapper[4771]: E1002 11:22:36.411342 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f334abc-1b22-4577-821b-ae4d10f2a1b6" containerName="extract-content" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.411348 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f334abc-1b22-4577-821b-ae4d10f2a1b6" containerName="extract-content" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.411564 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f334abc-1b22-4577-821b-ae4d10f2a1b6" containerName="registry-server" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.413665 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.428227 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-482nv"] Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.593436 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1e3530e-23b4-445b-97c9-f92613a54888-utilities\") pod \"redhat-marketplace-482nv\" (UID: \"b1e3530e-23b4-445b-97c9-f92613a54888\") " pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.593523 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx47f\" (UniqueName: \"kubernetes.io/projected/b1e3530e-23b4-445b-97c9-f92613a54888-kube-api-access-sx47f\") pod \"redhat-marketplace-482nv\" (UID: \"b1e3530e-23b4-445b-97c9-f92613a54888\") " pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.593674 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1e3530e-23b4-445b-97c9-f92613a54888-catalog-content\") pod \"redhat-marketplace-482nv\" (UID: \"b1e3530e-23b4-445b-97c9-f92613a54888\") " pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.695816 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1e3530e-23b4-445b-97c9-f92613a54888-utilities\") pod \"redhat-marketplace-482nv\" (UID: \"b1e3530e-23b4-445b-97c9-f92613a54888\") " pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.695866 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx47f\" (UniqueName: \"kubernetes.io/projected/b1e3530e-23b4-445b-97c9-f92613a54888-kube-api-access-sx47f\") pod \"redhat-marketplace-482nv\" (UID: \"b1e3530e-23b4-445b-97c9-f92613a54888\") " pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.696008 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1e3530e-23b4-445b-97c9-f92613a54888-catalog-content\") pod \"redhat-marketplace-482nv\" (UID: \"b1e3530e-23b4-445b-97c9-f92613a54888\") " pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.696506 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1e3530e-23b4-445b-97c9-f92613a54888-catalog-content\") pod \"redhat-marketplace-482nv\" (UID: \"b1e3530e-23b4-445b-97c9-f92613a54888\") " pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.696778 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1e3530e-23b4-445b-97c9-f92613a54888-utilities\") pod \"redhat-marketplace-482nv\" (UID: \"b1e3530e-23b4-445b-97c9-f92613a54888\") " pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.726020 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx47f\" (UniqueName: \"kubernetes.io/projected/b1e3530e-23b4-445b-97c9-f92613a54888-kube-api-access-sx47f\") pod \"redhat-marketplace-482nv\" (UID: \"b1e3530e-23b4-445b-97c9-f92613a54888\") " pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:36 crc kubenswrapper[4771]: I1002 11:22:36.736349 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:37 crc kubenswrapper[4771]: I1002 11:22:37.231996 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-482nv"] Oct 02 11:22:37 crc kubenswrapper[4771]: I1002 11:22:37.426232 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-482nv" event={"ID":"b1e3530e-23b4-445b-97c9-f92613a54888","Type":"ContainerStarted","Data":"798c1ac1fdd43d3b0790437569cc7935b8fe26ea2e41266eaddf8a41b5595722"} Oct 02 11:22:38 crc kubenswrapper[4771]: I1002 11:22:38.453385 4771 generic.go:334] "Generic (PLEG): container finished" podID="b1e3530e-23b4-445b-97c9-f92613a54888" containerID="602189addc3658fa88506390f738694482e03c567ff7870907aa2db174635314" exitCode=0 Oct 02 11:22:38 crc kubenswrapper[4771]: I1002 11:22:38.453431 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-482nv" event={"ID":"b1e3530e-23b4-445b-97c9-f92613a54888","Type":"ContainerDied","Data":"602189addc3658fa88506390f738694482e03c567ff7870907aa2db174635314"} Oct 02 11:22:40 crc kubenswrapper[4771]: E1002 11:22:40.351824 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1e3530e_23b4_445b_97c9_f92613a54888.slice/crio-conmon-fe12939ff6152179ed58bf1bfd9423c525a0f21290d13643e3bb6bba351f86ab.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1e3530e_23b4_445b_97c9_f92613a54888.slice/crio-fe12939ff6152179ed58bf1bfd9423c525a0f21290d13643e3bb6bba351f86ab.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:22:40 crc kubenswrapper[4771]: I1002 11:22:40.480928 4771 generic.go:334] "Generic (PLEG): container finished" podID="b1e3530e-23b4-445b-97c9-f92613a54888" containerID="fe12939ff6152179ed58bf1bfd9423c525a0f21290d13643e3bb6bba351f86ab" exitCode=0 Oct 02 11:22:40 crc kubenswrapper[4771]: I1002 11:22:40.480971 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-482nv" event={"ID":"b1e3530e-23b4-445b-97c9-f92613a54888","Type":"ContainerDied","Data":"fe12939ff6152179ed58bf1bfd9423c525a0f21290d13643e3bb6bba351f86ab"} Oct 02 11:22:41 crc kubenswrapper[4771]: I1002 11:22:41.497536 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-482nv" event={"ID":"b1e3530e-23b4-445b-97c9-f92613a54888","Type":"ContainerStarted","Data":"dbd80cd8ce64edc8b60ade5db9700ddcca9c29dea71a1968d913cecf6fb90eaa"} Oct 02 11:22:41 crc kubenswrapper[4771]: I1002 11:22:41.527505 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-482nv" podStartSLOduration=2.790240114 podStartE2EDuration="5.527487267s" podCreationTimestamp="2025-10-02 11:22:36 +0000 UTC" firstStartedPulling="2025-10-02 11:22:38.457644664 +0000 UTC m=+6346.105329731" lastFinishedPulling="2025-10-02 11:22:41.194891817 +0000 UTC m=+6348.842576884" observedRunningTime="2025-10-02 11:22:41.521984094 +0000 UTC m=+6349.169669161" watchObservedRunningTime="2025-10-02 11:22:41.527487267 +0000 UTC m=+6349.175172324" Oct 02 11:22:42 crc kubenswrapper[4771]: I1002 11:22:42.145855 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:22:42 crc kubenswrapper[4771]: I1002 11:22:42.146193 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:22:46 crc kubenswrapper[4771]: I1002 11:22:46.737036 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:46 crc kubenswrapper[4771]: I1002 11:22:46.738207 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:46 crc kubenswrapper[4771]: I1002 11:22:46.791413 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:47 crc kubenswrapper[4771]: I1002 11:22:47.623599 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:47 crc kubenswrapper[4771]: I1002 11:22:47.685053 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-482nv"] Oct 02 11:22:49 crc kubenswrapper[4771]: I1002 11:22:49.585335 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-482nv" podUID="b1e3530e-23b4-445b-97c9-f92613a54888" containerName="registry-server" containerID="cri-o://dbd80cd8ce64edc8b60ade5db9700ddcca9c29dea71a1968d913cecf6fb90eaa" gracePeriod=2 Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.177197 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.255160 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sx47f\" (UniqueName: \"kubernetes.io/projected/b1e3530e-23b4-445b-97c9-f92613a54888-kube-api-access-sx47f\") pod \"b1e3530e-23b4-445b-97c9-f92613a54888\" (UID: \"b1e3530e-23b4-445b-97c9-f92613a54888\") " Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.255708 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1e3530e-23b4-445b-97c9-f92613a54888-catalog-content\") pod \"b1e3530e-23b4-445b-97c9-f92613a54888\" (UID: \"b1e3530e-23b4-445b-97c9-f92613a54888\") " Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.255920 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1e3530e-23b4-445b-97c9-f92613a54888-utilities\") pod \"b1e3530e-23b4-445b-97c9-f92613a54888\" (UID: \"b1e3530e-23b4-445b-97c9-f92613a54888\") " Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.257714 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1e3530e-23b4-445b-97c9-f92613a54888-utilities" (OuterVolumeSpecName: "utilities") pod "b1e3530e-23b4-445b-97c9-f92613a54888" (UID: "b1e3530e-23b4-445b-97c9-f92613a54888"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.264450 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1e3530e-23b4-445b-97c9-f92613a54888-kube-api-access-sx47f" (OuterVolumeSpecName: "kube-api-access-sx47f") pod "b1e3530e-23b4-445b-97c9-f92613a54888" (UID: "b1e3530e-23b4-445b-97c9-f92613a54888"). InnerVolumeSpecName "kube-api-access-sx47f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.271804 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1e3530e-23b4-445b-97c9-f92613a54888-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1e3530e-23b4-445b-97c9-f92613a54888" (UID: "b1e3530e-23b4-445b-97c9-f92613a54888"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.358481 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1e3530e-23b4-445b-97c9-f92613a54888-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.358518 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1e3530e-23b4-445b-97c9-f92613a54888-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.358528 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx47f\" (UniqueName: \"kubernetes.io/projected/b1e3530e-23b4-445b-97c9-f92613a54888-kube-api-access-sx47f\") on node \"crc\" DevicePath \"\"" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.608593 4771 generic.go:334] "Generic (PLEG): container finished" podID="b1e3530e-23b4-445b-97c9-f92613a54888" containerID="dbd80cd8ce64edc8b60ade5db9700ddcca9c29dea71a1968d913cecf6fb90eaa" exitCode=0 Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.608671 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-482nv" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.608666 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-482nv" event={"ID":"b1e3530e-23b4-445b-97c9-f92613a54888","Type":"ContainerDied","Data":"dbd80cd8ce64edc8b60ade5db9700ddcca9c29dea71a1968d913cecf6fb90eaa"} Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.610394 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-482nv" event={"ID":"b1e3530e-23b4-445b-97c9-f92613a54888","Type":"ContainerDied","Data":"798c1ac1fdd43d3b0790437569cc7935b8fe26ea2e41266eaddf8a41b5595722"} Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.610416 4771 scope.go:117] "RemoveContainer" containerID="dbd80cd8ce64edc8b60ade5db9700ddcca9c29dea71a1968d913cecf6fb90eaa" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.651665 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-482nv"] Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.658116 4771 scope.go:117] "RemoveContainer" containerID="fe12939ff6152179ed58bf1bfd9423c525a0f21290d13643e3bb6bba351f86ab" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.668867 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-482nv"] Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.707964 4771 scope.go:117] "RemoveContainer" containerID="602189addc3658fa88506390f738694482e03c567ff7870907aa2db174635314" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.799981 4771 scope.go:117] "RemoveContainer" containerID="dbd80cd8ce64edc8b60ade5db9700ddcca9c29dea71a1968d913cecf6fb90eaa" Oct 02 11:22:50 crc kubenswrapper[4771]: E1002 11:22:50.800869 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbd80cd8ce64edc8b60ade5db9700ddcca9c29dea71a1968d913cecf6fb90eaa\": container with ID starting with dbd80cd8ce64edc8b60ade5db9700ddcca9c29dea71a1968d913cecf6fb90eaa not found: ID does not exist" containerID="dbd80cd8ce64edc8b60ade5db9700ddcca9c29dea71a1968d913cecf6fb90eaa" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.800986 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbd80cd8ce64edc8b60ade5db9700ddcca9c29dea71a1968d913cecf6fb90eaa"} err="failed to get container status \"dbd80cd8ce64edc8b60ade5db9700ddcca9c29dea71a1968d913cecf6fb90eaa\": rpc error: code = NotFound desc = could not find container \"dbd80cd8ce64edc8b60ade5db9700ddcca9c29dea71a1968d913cecf6fb90eaa\": container with ID starting with dbd80cd8ce64edc8b60ade5db9700ddcca9c29dea71a1968d913cecf6fb90eaa not found: ID does not exist" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.801117 4771 scope.go:117] "RemoveContainer" containerID="fe12939ff6152179ed58bf1bfd9423c525a0f21290d13643e3bb6bba351f86ab" Oct 02 11:22:50 crc kubenswrapper[4771]: E1002 11:22:50.801849 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe12939ff6152179ed58bf1bfd9423c525a0f21290d13643e3bb6bba351f86ab\": container with ID starting with fe12939ff6152179ed58bf1bfd9423c525a0f21290d13643e3bb6bba351f86ab not found: ID does not exist" containerID="fe12939ff6152179ed58bf1bfd9423c525a0f21290d13643e3bb6bba351f86ab" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.801952 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe12939ff6152179ed58bf1bfd9423c525a0f21290d13643e3bb6bba351f86ab"} err="failed to get container status \"fe12939ff6152179ed58bf1bfd9423c525a0f21290d13643e3bb6bba351f86ab\": rpc error: code = NotFound desc = could not find container \"fe12939ff6152179ed58bf1bfd9423c525a0f21290d13643e3bb6bba351f86ab\": container with ID starting with fe12939ff6152179ed58bf1bfd9423c525a0f21290d13643e3bb6bba351f86ab not found: ID does not exist" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.802058 4771 scope.go:117] "RemoveContainer" containerID="602189addc3658fa88506390f738694482e03c567ff7870907aa2db174635314" Oct 02 11:22:50 crc kubenswrapper[4771]: E1002 11:22:50.802863 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"602189addc3658fa88506390f738694482e03c567ff7870907aa2db174635314\": container with ID starting with 602189addc3658fa88506390f738694482e03c567ff7870907aa2db174635314 not found: ID does not exist" containerID="602189addc3658fa88506390f738694482e03c567ff7870907aa2db174635314" Oct 02 11:22:50 crc kubenswrapper[4771]: I1002 11:22:50.802906 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"602189addc3658fa88506390f738694482e03c567ff7870907aa2db174635314"} err="failed to get container status \"602189addc3658fa88506390f738694482e03c567ff7870907aa2db174635314\": rpc error: code = NotFound desc = could not find container \"602189addc3658fa88506390f738694482e03c567ff7870907aa2db174635314\": container with ID starting with 602189addc3658fa88506390f738694482e03c567ff7870907aa2db174635314 not found: ID does not exist" Oct 02 11:22:51 crc kubenswrapper[4771]: I1002 11:22:51.718931 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1e3530e-23b4-445b-97c9-f92613a54888" path="/var/lib/kubelet/pods/b1e3530e-23b4-445b-97c9-f92613a54888/volumes" Oct 02 11:23:12 crc kubenswrapper[4771]: I1002 11:23:12.146061 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:23:12 crc kubenswrapper[4771]: I1002 11:23:12.146611 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:23:42 crc kubenswrapper[4771]: I1002 11:23:42.145883 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:23:42 crc kubenswrapper[4771]: I1002 11:23:42.146593 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:23:42 crc kubenswrapper[4771]: I1002 11:23:42.146662 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 11:23:42 crc kubenswrapper[4771]: I1002 11:23:42.147572 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:23:42 crc kubenswrapper[4771]: I1002 11:23:42.147700 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" gracePeriod=600 Oct 02 11:23:42 crc kubenswrapper[4771]: E1002 11:23:42.287358 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:23:43 crc kubenswrapper[4771]: I1002 11:23:43.195898 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" exitCode=0 Oct 02 11:23:43 crc kubenswrapper[4771]: I1002 11:23:43.195967 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc"} Oct 02 11:23:43 crc kubenswrapper[4771]: I1002 11:23:43.196009 4771 scope.go:117] "RemoveContainer" containerID="5f3b64f0ba9c14fd906fe67d63409842d8b833922da63c8e0956c0158ecdb611" Oct 02 11:23:43 crc kubenswrapper[4771]: I1002 11:23:43.197308 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:23:43 crc kubenswrapper[4771]: E1002 11:23:43.197690 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:23:58 crc kubenswrapper[4771]: I1002 11:23:58.682353 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:23:58 crc kubenswrapper[4771]: E1002 11:23:58.683322 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:24:11 crc kubenswrapper[4771]: I1002 11:24:11.681255 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:24:11 crc kubenswrapper[4771]: E1002 11:24:11.682101 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:24:26 crc kubenswrapper[4771]: I1002 11:24:26.682054 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:24:26 crc kubenswrapper[4771]: E1002 11:24:26.684368 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:24:38 crc kubenswrapper[4771]: I1002 11:24:38.682421 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:24:38 crc kubenswrapper[4771]: E1002 11:24:38.683467 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:24:49 crc kubenswrapper[4771]: I1002 11:24:49.681987 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:24:49 crc kubenswrapper[4771]: E1002 11:24:49.682925 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:25:01 crc kubenswrapper[4771]: I1002 11:25:01.682684 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:25:01 crc kubenswrapper[4771]: E1002 11:25:01.684081 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:25:13 crc kubenswrapper[4771]: I1002 11:25:13.689595 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:25:13 crc kubenswrapper[4771]: E1002 11:25:13.690638 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:25:28 crc kubenswrapper[4771]: I1002 11:25:28.681812 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:25:28 crc kubenswrapper[4771]: E1002 11:25:28.682619 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:25:43 crc kubenswrapper[4771]: I1002 11:25:43.692857 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:25:43 crc kubenswrapper[4771]: E1002 11:25:43.694482 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:25:58 crc kubenswrapper[4771]: I1002 11:25:58.681290 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:25:58 crc kubenswrapper[4771]: E1002 11:25:58.682105 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.078912 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tkzlz"] Oct 02 11:26:05 crc kubenswrapper[4771]: E1002 11:26:05.080109 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e3530e-23b4-445b-97c9-f92613a54888" containerName="extract-content" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.080124 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e3530e-23b4-445b-97c9-f92613a54888" containerName="extract-content" Oct 02 11:26:05 crc kubenswrapper[4771]: E1002 11:26:05.080166 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e3530e-23b4-445b-97c9-f92613a54888" containerName="extract-utilities" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.080172 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e3530e-23b4-445b-97c9-f92613a54888" containerName="extract-utilities" Oct 02 11:26:05 crc kubenswrapper[4771]: E1002 11:26:05.080182 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e3530e-23b4-445b-97c9-f92613a54888" containerName="registry-server" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.080188 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e3530e-23b4-445b-97c9-f92613a54888" containerName="registry-server" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.080460 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e3530e-23b4-445b-97c9-f92613a54888" containerName="registry-server" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.082304 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.096705 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tkzlz"] Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.213377 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90fd4793-3fd1-4832-bf2c-e9f326ac2373-utilities\") pod \"redhat-operators-tkzlz\" (UID: \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\") " pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.213572 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90fd4793-3fd1-4832-bf2c-e9f326ac2373-catalog-content\") pod \"redhat-operators-tkzlz\" (UID: \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\") " pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.213823 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdd7p\" (UniqueName: \"kubernetes.io/projected/90fd4793-3fd1-4832-bf2c-e9f326ac2373-kube-api-access-jdd7p\") pod \"redhat-operators-tkzlz\" (UID: \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\") " pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.317287 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdd7p\" (UniqueName: \"kubernetes.io/projected/90fd4793-3fd1-4832-bf2c-e9f326ac2373-kube-api-access-jdd7p\") pod \"redhat-operators-tkzlz\" (UID: \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\") " pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.317439 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90fd4793-3fd1-4832-bf2c-e9f326ac2373-utilities\") pod \"redhat-operators-tkzlz\" (UID: \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\") " pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.317535 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90fd4793-3fd1-4832-bf2c-e9f326ac2373-catalog-content\") pod \"redhat-operators-tkzlz\" (UID: \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\") " pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.318144 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90fd4793-3fd1-4832-bf2c-e9f326ac2373-utilities\") pod \"redhat-operators-tkzlz\" (UID: \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\") " pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.318224 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90fd4793-3fd1-4832-bf2c-e9f326ac2373-catalog-content\") pod \"redhat-operators-tkzlz\" (UID: \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\") " pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.340368 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdd7p\" (UniqueName: \"kubernetes.io/projected/90fd4793-3fd1-4832-bf2c-e9f326ac2373-kube-api-access-jdd7p\") pod \"redhat-operators-tkzlz\" (UID: \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\") " pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.405762 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:05 crc kubenswrapper[4771]: I1002 11:26:05.889222 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tkzlz"] Oct 02 11:26:06 crc kubenswrapper[4771]: I1002 11:26:06.822689 4771 generic.go:334] "Generic (PLEG): container finished" podID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerID="28ff9001e630e4ff30b8b686dbef6e8ca704d0ca46ddc071371d1e16bb1e0e8b" exitCode=0 Oct 02 11:26:06 crc kubenswrapper[4771]: I1002 11:26:06.822810 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tkzlz" event={"ID":"90fd4793-3fd1-4832-bf2c-e9f326ac2373","Type":"ContainerDied","Data":"28ff9001e630e4ff30b8b686dbef6e8ca704d0ca46ddc071371d1e16bb1e0e8b"} Oct 02 11:26:06 crc kubenswrapper[4771]: I1002 11:26:06.823962 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tkzlz" event={"ID":"90fd4793-3fd1-4832-bf2c-e9f326ac2373","Type":"ContainerStarted","Data":"2417882d821574986b290936ad58f3ebebc8691af16821f54c18feaecb1f1244"} Oct 02 11:26:08 crc kubenswrapper[4771]: I1002 11:26:08.853775 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tkzlz" event={"ID":"90fd4793-3fd1-4832-bf2c-e9f326ac2373","Type":"ContainerStarted","Data":"bebe3b70ff3f53e2a9365f6e2ad4802abc538a643e4f1f42c30d4e525dcb900f"} Oct 02 11:26:10 crc kubenswrapper[4771]: I1002 11:26:10.682353 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:26:10 crc kubenswrapper[4771]: E1002 11:26:10.683884 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:26:12 crc kubenswrapper[4771]: I1002 11:26:12.894219 4771 generic.go:334] "Generic (PLEG): container finished" podID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerID="bebe3b70ff3f53e2a9365f6e2ad4802abc538a643e4f1f42c30d4e525dcb900f" exitCode=0 Oct 02 11:26:12 crc kubenswrapper[4771]: I1002 11:26:12.894316 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tkzlz" event={"ID":"90fd4793-3fd1-4832-bf2c-e9f326ac2373","Type":"ContainerDied","Data":"bebe3b70ff3f53e2a9365f6e2ad4802abc538a643e4f1f42c30d4e525dcb900f"} Oct 02 11:26:14 crc kubenswrapper[4771]: I1002 11:26:14.918281 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tkzlz" event={"ID":"90fd4793-3fd1-4832-bf2c-e9f326ac2373","Type":"ContainerStarted","Data":"5ff8da4635795f2b0e2b15de11e34ac83cafa49207a849f72c37ce460b9fe9ad"} Oct 02 11:26:14 crc kubenswrapper[4771]: I1002 11:26:14.936175 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tkzlz" podStartSLOduration=3.051068434 podStartE2EDuration="9.936159949s" podCreationTimestamp="2025-10-02 11:26:05 +0000 UTC" firstStartedPulling="2025-10-02 11:26:06.825443434 +0000 UTC m=+6554.473128501" lastFinishedPulling="2025-10-02 11:26:13.710534949 +0000 UTC m=+6561.358220016" observedRunningTime="2025-10-02 11:26:14.93465957 +0000 UTC m=+6562.582344637" watchObservedRunningTime="2025-10-02 11:26:14.936159949 +0000 UTC m=+6562.583845016" Oct 02 11:26:15 crc kubenswrapper[4771]: I1002 11:26:15.406584 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:15 crc kubenswrapper[4771]: I1002 11:26:15.406637 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:16 crc kubenswrapper[4771]: I1002 11:26:16.462383 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tkzlz" podUID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerName="registry-server" probeResult="failure" output=< Oct 02 11:26:16 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:26:16 crc kubenswrapper[4771]: > Oct 02 11:26:22 crc kubenswrapper[4771]: I1002 11:26:22.692873 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:26:22 crc kubenswrapper[4771]: E1002 11:26:22.696511 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:26:26 crc kubenswrapper[4771]: I1002 11:26:26.453426 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tkzlz" podUID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerName="registry-server" probeResult="failure" output=< Oct 02 11:26:26 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:26:26 crc kubenswrapper[4771]: > Oct 02 11:26:35 crc kubenswrapper[4771]: I1002 11:26:35.684440 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:26:35 crc kubenswrapper[4771]: E1002 11:26:35.685713 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:26:36 crc kubenswrapper[4771]: I1002 11:26:36.462330 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tkzlz" podUID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerName="registry-server" probeResult="failure" output=< Oct 02 11:26:36 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:26:36 crc kubenswrapper[4771]: > Oct 02 11:26:45 crc kubenswrapper[4771]: I1002 11:26:45.457040 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:45 crc kubenswrapper[4771]: I1002 11:26:45.521349 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:45 crc kubenswrapper[4771]: I1002 11:26:45.700900 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tkzlz"] Oct 02 11:26:47 crc kubenswrapper[4771]: I1002 11:26:47.318538 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tkzlz" podUID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerName="registry-server" containerID="cri-o://5ff8da4635795f2b0e2b15de11e34ac83cafa49207a849f72c37ce460b9fe9ad" gracePeriod=2 Oct 02 11:26:47 crc kubenswrapper[4771]: I1002 11:26:47.833977 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:47 crc kubenswrapper[4771]: I1002 11:26:47.883164 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90fd4793-3fd1-4832-bf2c-e9f326ac2373-utilities\") pod \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\" (UID: \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\") " Oct 02 11:26:47 crc kubenswrapper[4771]: I1002 11:26:47.883324 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdd7p\" (UniqueName: \"kubernetes.io/projected/90fd4793-3fd1-4832-bf2c-e9f326ac2373-kube-api-access-jdd7p\") pod \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\" (UID: \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\") " Oct 02 11:26:47 crc kubenswrapper[4771]: I1002 11:26:47.883367 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90fd4793-3fd1-4832-bf2c-e9f326ac2373-catalog-content\") pod \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\" (UID: \"90fd4793-3fd1-4832-bf2c-e9f326ac2373\") " Oct 02 11:26:47 crc kubenswrapper[4771]: I1002 11:26:47.884092 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90fd4793-3fd1-4832-bf2c-e9f326ac2373-utilities" (OuterVolumeSpecName: "utilities") pod "90fd4793-3fd1-4832-bf2c-e9f326ac2373" (UID: "90fd4793-3fd1-4832-bf2c-e9f326ac2373"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:26:47 crc kubenswrapper[4771]: I1002 11:26:47.890058 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90fd4793-3fd1-4832-bf2c-e9f326ac2373-kube-api-access-jdd7p" (OuterVolumeSpecName: "kube-api-access-jdd7p") pod "90fd4793-3fd1-4832-bf2c-e9f326ac2373" (UID: "90fd4793-3fd1-4832-bf2c-e9f326ac2373"). InnerVolumeSpecName "kube-api-access-jdd7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:26:47 crc kubenswrapper[4771]: I1002 11:26:47.985792 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90fd4793-3fd1-4832-bf2c-e9f326ac2373-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:26:47 crc kubenswrapper[4771]: I1002 11:26:47.985826 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdd7p\" (UniqueName: \"kubernetes.io/projected/90fd4793-3fd1-4832-bf2c-e9f326ac2373-kube-api-access-jdd7p\") on node \"crc\" DevicePath \"\"" Oct 02 11:26:47 crc kubenswrapper[4771]: I1002 11:26:47.985957 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90fd4793-3fd1-4832-bf2c-e9f326ac2373-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90fd4793-3fd1-4832-bf2c-e9f326ac2373" (UID: "90fd4793-3fd1-4832-bf2c-e9f326ac2373"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.088370 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90fd4793-3fd1-4832-bf2c-e9f326ac2373-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.370330 4771 generic.go:334] "Generic (PLEG): container finished" podID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerID="5ff8da4635795f2b0e2b15de11e34ac83cafa49207a849f72c37ce460b9fe9ad" exitCode=0 Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.371238 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tkzlz" Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.371881 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tkzlz" event={"ID":"90fd4793-3fd1-4832-bf2c-e9f326ac2373","Type":"ContainerDied","Data":"5ff8da4635795f2b0e2b15de11e34ac83cafa49207a849f72c37ce460b9fe9ad"} Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.371940 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tkzlz" event={"ID":"90fd4793-3fd1-4832-bf2c-e9f326ac2373","Type":"ContainerDied","Data":"2417882d821574986b290936ad58f3ebebc8691af16821f54c18feaecb1f1244"} Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.371961 4771 scope.go:117] "RemoveContainer" containerID="5ff8da4635795f2b0e2b15de11e34ac83cafa49207a849f72c37ce460b9fe9ad" Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.452233 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tkzlz"] Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.462768 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tkzlz"] Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.476352 4771 scope.go:117] "RemoveContainer" containerID="bebe3b70ff3f53e2a9365f6e2ad4802abc538a643e4f1f42c30d4e525dcb900f" Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.533989 4771 scope.go:117] "RemoveContainer" containerID="28ff9001e630e4ff30b8b686dbef6e8ca704d0ca46ddc071371d1e16bb1e0e8b" Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.608255 4771 scope.go:117] "RemoveContainer" containerID="5ff8da4635795f2b0e2b15de11e34ac83cafa49207a849f72c37ce460b9fe9ad" Oct 02 11:26:48 crc kubenswrapper[4771]: E1002 11:26:48.608726 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ff8da4635795f2b0e2b15de11e34ac83cafa49207a849f72c37ce460b9fe9ad\": container with ID starting with 5ff8da4635795f2b0e2b15de11e34ac83cafa49207a849f72c37ce460b9fe9ad not found: ID does not exist" containerID="5ff8da4635795f2b0e2b15de11e34ac83cafa49207a849f72c37ce460b9fe9ad" Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.608788 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ff8da4635795f2b0e2b15de11e34ac83cafa49207a849f72c37ce460b9fe9ad"} err="failed to get container status \"5ff8da4635795f2b0e2b15de11e34ac83cafa49207a849f72c37ce460b9fe9ad\": rpc error: code = NotFound desc = could not find container \"5ff8da4635795f2b0e2b15de11e34ac83cafa49207a849f72c37ce460b9fe9ad\": container with ID starting with 5ff8da4635795f2b0e2b15de11e34ac83cafa49207a849f72c37ce460b9fe9ad not found: ID does not exist" Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.608814 4771 scope.go:117] "RemoveContainer" containerID="bebe3b70ff3f53e2a9365f6e2ad4802abc538a643e4f1f42c30d4e525dcb900f" Oct 02 11:26:48 crc kubenswrapper[4771]: E1002 11:26:48.609454 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bebe3b70ff3f53e2a9365f6e2ad4802abc538a643e4f1f42c30d4e525dcb900f\": container with ID starting with bebe3b70ff3f53e2a9365f6e2ad4802abc538a643e4f1f42c30d4e525dcb900f not found: ID does not exist" containerID="bebe3b70ff3f53e2a9365f6e2ad4802abc538a643e4f1f42c30d4e525dcb900f" Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.609482 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bebe3b70ff3f53e2a9365f6e2ad4802abc538a643e4f1f42c30d4e525dcb900f"} err="failed to get container status \"bebe3b70ff3f53e2a9365f6e2ad4802abc538a643e4f1f42c30d4e525dcb900f\": rpc error: code = NotFound desc = could not find container \"bebe3b70ff3f53e2a9365f6e2ad4802abc538a643e4f1f42c30d4e525dcb900f\": container with ID starting with bebe3b70ff3f53e2a9365f6e2ad4802abc538a643e4f1f42c30d4e525dcb900f not found: ID does not exist" Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.609502 4771 scope.go:117] "RemoveContainer" containerID="28ff9001e630e4ff30b8b686dbef6e8ca704d0ca46ddc071371d1e16bb1e0e8b" Oct 02 11:26:48 crc kubenswrapper[4771]: E1002 11:26:48.609749 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28ff9001e630e4ff30b8b686dbef6e8ca704d0ca46ddc071371d1e16bb1e0e8b\": container with ID starting with 28ff9001e630e4ff30b8b686dbef6e8ca704d0ca46ddc071371d1e16bb1e0e8b not found: ID does not exist" containerID="28ff9001e630e4ff30b8b686dbef6e8ca704d0ca46ddc071371d1e16bb1e0e8b" Oct 02 11:26:48 crc kubenswrapper[4771]: I1002 11:26:48.609771 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28ff9001e630e4ff30b8b686dbef6e8ca704d0ca46ddc071371d1e16bb1e0e8b"} err="failed to get container status \"28ff9001e630e4ff30b8b686dbef6e8ca704d0ca46ddc071371d1e16bb1e0e8b\": rpc error: code = NotFound desc = could not find container \"28ff9001e630e4ff30b8b686dbef6e8ca704d0ca46ddc071371d1e16bb1e0e8b\": container with ID starting with 28ff9001e630e4ff30b8b686dbef6e8ca704d0ca46ddc071371d1e16bb1e0e8b not found: ID does not exist" Oct 02 11:26:49 crc kubenswrapper[4771]: I1002 11:26:49.696883 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" path="/var/lib/kubelet/pods/90fd4793-3fd1-4832-bf2c-e9f326ac2373/volumes" Oct 02 11:26:50 crc kubenswrapper[4771]: I1002 11:26:50.681647 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:26:50 crc kubenswrapper[4771]: E1002 11:26:50.682375 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:27:01 crc kubenswrapper[4771]: I1002 11:27:01.682184 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:27:01 crc kubenswrapper[4771]: E1002 11:27:01.682948 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.466146 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lskfb"] Oct 02 11:27:16 crc kubenswrapper[4771]: E1002 11:27:16.467195 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerName="extract-utilities" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.467210 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerName="extract-utilities" Oct 02 11:27:16 crc kubenswrapper[4771]: E1002 11:27:16.467232 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerName="registry-server" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.467239 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerName="registry-server" Oct 02 11:27:16 crc kubenswrapper[4771]: E1002 11:27:16.467265 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerName="extract-content" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.467273 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerName="extract-content" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.467544 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="90fd4793-3fd1-4832-bf2c-e9f326ac2373" containerName="registry-server" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.469463 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.483246 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lskfb"] Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.539280 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-utilities\") pod \"community-operators-lskfb\" (UID: \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\") " pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.539613 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-catalog-content\") pod \"community-operators-lskfb\" (UID: \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\") " pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.539693 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbw8r\" (UniqueName: \"kubernetes.io/projected/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-kube-api-access-qbw8r\") pod \"community-operators-lskfb\" (UID: \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\") " pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.642516 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-catalog-content\") pod \"community-operators-lskfb\" (UID: \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\") " pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.642566 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbw8r\" (UniqueName: \"kubernetes.io/projected/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-kube-api-access-qbw8r\") pod \"community-operators-lskfb\" (UID: \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\") " pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.642716 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-utilities\") pod \"community-operators-lskfb\" (UID: \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\") " pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.643054 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-catalog-content\") pod \"community-operators-lskfb\" (UID: \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\") " pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.643244 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-utilities\") pod \"community-operators-lskfb\" (UID: \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\") " pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.669335 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbw8r\" (UniqueName: \"kubernetes.io/projected/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-kube-api-access-qbw8r\") pod \"community-operators-lskfb\" (UID: \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\") " pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.681799 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:27:16 crc kubenswrapper[4771]: E1002 11:27:16.682227 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:27:16 crc kubenswrapper[4771]: I1002 11:27:16.806409 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:17 crc kubenswrapper[4771]: I1002 11:27:17.439905 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lskfb"] Oct 02 11:27:17 crc kubenswrapper[4771]: I1002 11:27:17.745872 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lskfb" event={"ID":"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48","Type":"ContainerStarted","Data":"cc901b9caeccb33d76a30e42340cba655f45b4f0467537c291c5429964b8845d"} Oct 02 11:27:18 crc kubenswrapper[4771]: I1002 11:27:18.758698 4771 generic.go:334] "Generic (PLEG): container finished" podID="a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" containerID="a8eae3858d3a2694462f3d5627879852068b74b93760194e440341c4f9459388" exitCode=0 Oct 02 11:27:18 crc kubenswrapper[4771]: I1002 11:27:18.758747 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lskfb" event={"ID":"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48","Type":"ContainerDied","Data":"a8eae3858d3a2694462f3d5627879852068b74b93760194e440341c4f9459388"} Oct 02 11:27:18 crc kubenswrapper[4771]: I1002 11:27:18.761397 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:27:20 crc kubenswrapper[4771]: I1002 11:27:20.781437 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lskfb" event={"ID":"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48","Type":"ContainerStarted","Data":"53984477822ddda9cef70d68a8bed3f92ff235596fa44ae61469c4c3131c0ae8"} Oct 02 11:27:23 crc kubenswrapper[4771]: I1002 11:27:23.815654 4771 generic.go:334] "Generic (PLEG): container finished" podID="a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" containerID="53984477822ddda9cef70d68a8bed3f92ff235596fa44ae61469c4c3131c0ae8" exitCode=0 Oct 02 11:27:23 crc kubenswrapper[4771]: I1002 11:27:23.816295 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lskfb" event={"ID":"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48","Type":"ContainerDied","Data":"53984477822ddda9cef70d68a8bed3f92ff235596fa44ae61469c4c3131c0ae8"} Oct 02 11:27:24 crc kubenswrapper[4771]: I1002 11:27:24.828436 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lskfb" event={"ID":"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48","Type":"ContainerStarted","Data":"bff2a45069425188ef18c709ca20ef59d4241b31e6854567f28c6777618d5946"} Oct 02 11:27:24 crc kubenswrapper[4771]: I1002 11:27:24.860486 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lskfb" podStartSLOduration=3.165510421 podStartE2EDuration="8.860469514s" podCreationTimestamp="2025-10-02 11:27:16 +0000 UTC" firstStartedPulling="2025-10-02 11:27:18.761149482 +0000 UTC m=+6626.408834549" lastFinishedPulling="2025-10-02 11:27:24.456108575 +0000 UTC m=+6632.103793642" observedRunningTime="2025-10-02 11:27:24.845325711 +0000 UTC m=+6632.493010768" watchObservedRunningTime="2025-10-02 11:27:24.860469514 +0000 UTC m=+6632.508154571" Oct 02 11:27:26 crc kubenswrapper[4771]: I1002 11:27:26.806719 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:26 crc kubenswrapper[4771]: I1002 11:27:26.807180 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:26 crc kubenswrapper[4771]: I1002 11:27:26.859416 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:30 crc kubenswrapper[4771]: I1002 11:27:30.681194 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:27:30 crc kubenswrapper[4771]: E1002 11:27:30.681779 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:27:36 crc kubenswrapper[4771]: I1002 11:27:36.860701 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:36 crc kubenswrapper[4771]: I1002 11:27:36.924168 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lskfb"] Oct 02 11:27:36 crc kubenswrapper[4771]: I1002 11:27:36.960631 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lskfb" podUID="a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" containerName="registry-server" containerID="cri-o://bff2a45069425188ef18c709ca20ef59d4241b31e6854567f28c6777618d5946" gracePeriod=2 Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.529425 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.577059 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-utilities\") pod \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\" (UID: \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\") " Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.577466 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-catalog-content\") pod \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\" (UID: \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\") " Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.577557 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbw8r\" (UniqueName: \"kubernetes.io/projected/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-kube-api-access-qbw8r\") pod \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\" (UID: \"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48\") " Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.586212 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-kube-api-access-qbw8r" (OuterVolumeSpecName: "kube-api-access-qbw8r") pod "a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" (UID: "a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48"). InnerVolumeSpecName "kube-api-access-qbw8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.587074 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-utilities" (OuterVolumeSpecName: "utilities") pod "a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" (UID: "a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.648350 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" (UID: "a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.680199 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.680230 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.680241 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbw8r\" (UniqueName: \"kubernetes.io/projected/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48-kube-api-access-qbw8r\") on node \"crc\" DevicePath \"\"" Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.977701 4771 generic.go:334] "Generic (PLEG): container finished" podID="a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" containerID="bff2a45069425188ef18c709ca20ef59d4241b31e6854567f28c6777618d5946" exitCode=0 Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.977745 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lskfb" event={"ID":"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48","Type":"ContainerDied","Data":"bff2a45069425188ef18c709ca20ef59d4241b31e6854567f28c6777618d5946"} Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.977774 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lskfb" event={"ID":"a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48","Type":"ContainerDied","Data":"cc901b9caeccb33d76a30e42340cba655f45b4f0467537c291c5429964b8845d"} Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.977798 4771 scope.go:117] "RemoveContainer" containerID="bff2a45069425188ef18c709ca20ef59d4241b31e6854567f28c6777618d5946" Oct 02 11:27:37 crc kubenswrapper[4771]: I1002 11:27:37.977797 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lskfb" Oct 02 11:27:38 crc kubenswrapper[4771]: I1002 11:27:38.025977 4771 scope.go:117] "RemoveContainer" containerID="53984477822ddda9cef70d68a8bed3f92ff235596fa44ae61469c4c3131c0ae8" Oct 02 11:27:38 crc kubenswrapper[4771]: I1002 11:27:38.033285 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lskfb"] Oct 02 11:27:38 crc kubenswrapper[4771]: I1002 11:27:38.047181 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lskfb"] Oct 02 11:27:38 crc kubenswrapper[4771]: I1002 11:27:38.055284 4771 scope.go:117] "RemoveContainer" containerID="a8eae3858d3a2694462f3d5627879852068b74b93760194e440341c4f9459388" Oct 02 11:27:38 crc kubenswrapper[4771]: I1002 11:27:38.116352 4771 scope.go:117] "RemoveContainer" containerID="bff2a45069425188ef18c709ca20ef59d4241b31e6854567f28c6777618d5946" Oct 02 11:27:38 crc kubenswrapper[4771]: E1002 11:27:38.117687 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bff2a45069425188ef18c709ca20ef59d4241b31e6854567f28c6777618d5946\": container with ID starting with bff2a45069425188ef18c709ca20ef59d4241b31e6854567f28c6777618d5946 not found: ID does not exist" containerID="bff2a45069425188ef18c709ca20ef59d4241b31e6854567f28c6777618d5946" Oct 02 11:27:38 crc kubenswrapper[4771]: I1002 11:27:38.117730 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bff2a45069425188ef18c709ca20ef59d4241b31e6854567f28c6777618d5946"} err="failed to get container status \"bff2a45069425188ef18c709ca20ef59d4241b31e6854567f28c6777618d5946\": rpc error: code = NotFound desc = could not find container \"bff2a45069425188ef18c709ca20ef59d4241b31e6854567f28c6777618d5946\": container with ID starting with bff2a45069425188ef18c709ca20ef59d4241b31e6854567f28c6777618d5946 not found: ID does not exist" Oct 02 11:27:38 crc kubenswrapper[4771]: I1002 11:27:38.117756 4771 scope.go:117] "RemoveContainer" containerID="53984477822ddda9cef70d68a8bed3f92ff235596fa44ae61469c4c3131c0ae8" Oct 02 11:27:38 crc kubenswrapper[4771]: E1002 11:27:38.118093 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53984477822ddda9cef70d68a8bed3f92ff235596fa44ae61469c4c3131c0ae8\": container with ID starting with 53984477822ddda9cef70d68a8bed3f92ff235596fa44ae61469c4c3131c0ae8 not found: ID does not exist" containerID="53984477822ddda9cef70d68a8bed3f92ff235596fa44ae61469c4c3131c0ae8" Oct 02 11:27:38 crc kubenswrapper[4771]: I1002 11:27:38.118138 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53984477822ddda9cef70d68a8bed3f92ff235596fa44ae61469c4c3131c0ae8"} err="failed to get container status \"53984477822ddda9cef70d68a8bed3f92ff235596fa44ae61469c4c3131c0ae8\": rpc error: code = NotFound desc = could not find container \"53984477822ddda9cef70d68a8bed3f92ff235596fa44ae61469c4c3131c0ae8\": container with ID starting with 53984477822ddda9cef70d68a8bed3f92ff235596fa44ae61469c4c3131c0ae8 not found: ID does not exist" Oct 02 11:27:38 crc kubenswrapper[4771]: I1002 11:27:38.118157 4771 scope.go:117] "RemoveContainer" containerID="a8eae3858d3a2694462f3d5627879852068b74b93760194e440341c4f9459388" Oct 02 11:27:38 crc kubenswrapper[4771]: E1002 11:27:38.118390 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8eae3858d3a2694462f3d5627879852068b74b93760194e440341c4f9459388\": container with ID starting with a8eae3858d3a2694462f3d5627879852068b74b93760194e440341c4f9459388 not found: ID does not exist" containerID="a8eae3858d3a2694462f3d5627879852068b74b93760194e440341c4f9459388" Oct 02 11:27:38 crc kubenswrapper[4771]: I1002 11:27:38.118416 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8eae3858d3a2694462f3d5627879852068b74b93760194e440341c4f9459388"} err="failed to get container status \"a8eae3858d3a2694462f3d5627879852068b74b93760194e440341c4f9459388\": rpc error: code = NotFound desc = could not find container \"a8eae3858d3a2694462f3d5627879852068b74b93760194e440341c4f9459388\": container with ID starting with a8eae3858d3a2694462f3d5627879852068b74b93760194e440341c4f9459388 not found: ID does not exist" Oct 02 11:27:39 crc kubenswrapper[4771]: I1002 11:27:39.696865 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" path="/var/lib/kubelet/pods/a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48/volumes" Oct 02 11:27:45 crc kubenswrapper[4771]: I1002 11:27:45.681370 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:27:45 crc kubenswrapper[4771]: E1002 11:27:45.682292 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:27:58 crc kubenswrapper[4771]: I1002 11:27:58.685825 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:27:58 crc kubenswrapper[4771]: E1002 11:27:58.686900 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:28:09 crc kubenswrapper[4771]: I1002 11:28:09.681981 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:28:09 crc kubenswrapper[4771]: E1002 11:28:09.682958 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:28:22 crc kubenswrapper[4771]: I1002 11:28:22.681442 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:28:22 crc kubenswrapper[4771]: E1002 11:28:22.683232 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:28:37 crc kubenswrapper[4771]: I1002 11:28:37.681981 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:28:37 crc kubenswrapper[4771]: E1002 11:28:37.682793 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:28:51 crc kubenswrapper[4771]: I1002 11:28:51.682889 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:28:52 crc kubenswrapper[4771]: I1002 11:28:52.805059 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"2486e5a32ac351379d45c0c9b8440f7fe86fe9b348bd28df0df4a5133b9cb5b8"} Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.160110 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk"] Oct 02 11:30:00 crc kubenswrapper[4771]: E1002 11:30:00.161224 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" containerName="registry-server" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.161242 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" containerName="registry-server" Oct 02 11:30:00 crc kubenswrapper[4771]: E1002 11:30:00.161296 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" containerName="extract-utilities" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.161303 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" containerName="extract-utilities" Oct 02 11:30:00 crc kubenswrapper[4771]: E1002 11:30:00.161337 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" containerName="extract-content" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.161345 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" containerName="extract-content" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.161638 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a29d8ad3-e51a-4a9a-b5ad-3ce140afdc48" containerName="registry-server" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.162704 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.165375 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.165459 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.172891 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk"] Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.266840 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-secret-volume\") pod \"collect-profiles-29323410-5v4hk\" (UID: \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.266900 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8p4n\" (UniqueName: \"kubernetes.io/projected/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-kube-api-access-p8p4n\") pod \"collect-profiles-29323410-5v4hk\" (UID: \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.266924 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-config-volume\") pod \"collect-profiles-29323410-5v4hk\" (UID: \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.369375 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-secret-volume\") pod \"collect-profiles-29323410-5v4hk\" (UID: \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.369431 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8p4n\" (UniqueName: \"kubernetes.io/projected/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-kube-api-access-p8p4n\") pod \"collect-profiles-29323410-5v4hk\" (UID: \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.369458 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-config-volume\") pod \"collect-profiles-29323410-5v4hk\" (UID: \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.370719 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-config-volume\") pod \"collect-profiles-29323410-5v4hk\" (UID: \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.375686 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-secret-volume\") pod \"collect-profiles-29323410-5v4hk\" (UID: \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.390269 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8p4n\" (UniqueName: \"kubernetes.io/projected/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-kube-api-access-p8p4n\") pod \"collect-profiles-29323410-5v4hk\" (UID: \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" Oct 02 11:30:00 crc kubenswrapper[4771]: I1002 11:30:00.490185 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" Oct 02 11:30:01 crc kubenswrapper[4771]: I1002 11:30:01.052364 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk"] Oct 02 11:30:01 crc kubenswrapper[4771]: I1002 11:30:01.564079 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" event={"ID":"aad1bd0b-36ce-4f6e-9ef0-261b539d9187","Type":"ContainerStarted","Data":"476ee740dccc9ac951cb9f11b18dcf82a2b43c93b45f2c3493f4564908514105"} Oct 02 11:30:01 crc kubenswrapper[4771]: I1002 11:30:01.564224 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" event={"ID":"aad1bd0b-36ce-4f6e-9ef0-261b539d9187","Type":"ContainerStarted","Data":"9b0d30390d02745a8cbf17249f6d0a8f5f71b8f8934ba34e542a567e0c4d9e51"} Oct 02 11:30:01 crc kubenswrapper[4771]: I1002 11:30:01.593613 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" podStartSLOduration=1.593579788 podStartE2EDuration="1.593579788s" podCreationTimestamp="2025-10-02 11:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:30:01.583103577 +0000 UTC m=+6789.230788644" watchObservedRunningTime="2025-10-02 11:30:01.593579788 +0000 UTC m=+6789.241264845" Oct 02 11:30:02 crc kubenswrapper[4771]: I1002 11:30:02.576951 4771 generic.go:334] "Generic (PLEG): container finished" podID="aad1bd0b-36ce-4f6e-9ef0-261b539d9187" containerID="476ee740dccc9ac951cb9f11b18dcf82a2b43c93b45f2c3493f4564908514105" exitCode=0 Oct 02 11:30:02 crc kubenswrapper[4771]: I1002 11:30:02.577009 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" event={"ID":"aad1bd0b-36ce-4f6e-9ef0-261b539d9187","Type":"ContainerDied","Data":"476ee740dccc9ac951cb9f11b18dcf82a2b43c93b45f2c3493f4564908514105"} Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.058088 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.208043 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-config-volume\") pod \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\" (UID: \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\") " Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.208354 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8p4n\" (UniqueName: \"kubernetes.io/projected/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-kube-api-access-p8p4n\") pod \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\" (UID: \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\") " Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.208394 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-secret-volume\") pod \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\" (UID: \"aad1bd0b-36ce-4f6e-9ef0-261b539d9187\") " Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.208818 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-config-volume" (OuterVolumeSpecName: "config-volume") pod "aad1bd0b-36ce-4f6e-9ef0-261b539d9187" (UID: "aad1bd0b-36ce-4f6e-9ef0-261b539d9187"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.209158 4771 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.215007 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "aad1bd0b-36ce-4f6e-9ef0-261b539d9187" (UID: "aad1bd0b-36ce-4f6e-9ef0-261b539d9187"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.215424 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-kube-api-access-p8p4n" (OuterVolumeSpecName: "kube-api-access-p8p4n") pod "aad1bd0b-36ce-4f6e-9ef0-261b539d9187" (UID: "aad1bd0b-36ce-4f6e-9ef0-261b539d9187"). InnerVolumeSpecName "kube-api-access-p8p4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.311036 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8p4n\" (UniqueName: \"kubernetes.io/projected/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-kube-api-access-p8p4n\") on node \"crc\" DevicePath \"\"" Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.311077 4771 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aad1bd0b-36ce-4f6e-9ef0-261b539d9187-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.599466 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" event={"ID":"aad1bd0b-36ce-4f6e-9ef0-261b539d9187","Type":"ContainerDied","Data":"9b0d30390d02745a8cbf17249f6d0a8f5f71b8f8934ba34e542a567e0c4d9e51"} Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.599757 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b0d30390d02745a8cbf17249f6d0a8f5f71b8f8934ba34e542a567e0c4d9e51" Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.599647 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5v4hk" Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.659454 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg"] Oct 02 11:30:04 crc kubenswrapper[4771]: I1002 11:30:04.669002 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323365-8nwzg"] Oct 02 11:30:05 crc kubenswrapper[4771]: I1002 11:30:05.711362 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ee7e9cd-4105-4ea8-bdfc-ab469fa68788" path="/var/lib/kubelet/pods/0ee7e9cd-4105-4ea8-bdfc-ab469fa68788/volumes" Oct 02 11:30:07 crc kubenswrapper[4771]: I1002 11:30:07.203002 4771 scope.go:117] "RemoveContainer" containerID="1a064682a845bcc5e3e48492a585aa00f2ce166c5bfdcd5315b2b3e4b2312760" Oct 02 11:31:12 crc kubenswrapper[4771]: I1002 11:31:12.146024 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:31:12 crc kubenswrapper[4771]: I1002 11:31:12.146651 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:31:42 crc kubenswrapper[4771]: I1002 11:31:42.145806 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:31:42 crc kubenswrapper[4771]: I1002 11:31:42.146339 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:31:56 crc kubenswrapper[4771]: I1002 11:31:56.957001 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xlvnh"] Oct 02 11:31:56 crc kubenswrapper[4771]: E1002 11:31:56.958121 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad1bd0b-36ce-4f6e-9ef0-261b539d9187" containerName="collect-profiles" Oct 02 11:31:56 crc kubenswrapper[4771]: I1002 11:31:56.958152 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad1bd0b-36ce-4f6e-9ef0-261b539d9187" containerName="collect-profiles" Oct 02 11:31:56 crc kubenswrapper[4771]: I1002 11:31:56.958384 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad1bd0b-36ce-4f6e-9ef0-261b539d9187" containerName="collect-profiles" Oct 02 11:31:56 crc kubenswrapper[4771]: I1002 11:31:56.960177 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:31:56 crc kubenswrapper[4771]: I1002 11:31:56.969078 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xlvnh"] Oct 02 11:31:57 crc kubenswrapper[4771]: I1002 11:31:57.072279 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c9d376-d940-4f03-84b0-318e08394c5f-catalog-content\") pod \"certified-operators-xlvnh\" (UID: \"d5c9d376-d940-4f03-84b0-318e08394c5f\") " pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:31:57 crc kubenswrapper[4771]: I1002 11:31:57.072788 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2hhg\" (UniqueName: \"kubernetes.io/projected/d5c9d376-d940-4f03-84b0-318e08394c5f-kube-api-access-k2hhg\") pod \"certified-operators-xlvnh\" (UID: \"d5c9d376-d940-4f03-84b0-318e08394c5f\") " pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:31:57 crc kubenswrapper[4771]: I1002 11:31:57.072848 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c9d376-d940-4f03-84b0-318e08394c5f-utilities\") pod \"certified-operators-xlvnh\" (UID: \"d5c9d376-d940-4f03-84b0-318e08394c5f\") " pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:31:57 crc kubenswrapper[4771]: I1002 11:31:57.175353 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c9d376-d940-4f03-84b0-318e08394c5f-catalog-content\") pod \"certified-operators-xlvnh\" (UID: \"d5c9d376-d940-4f03-84b0-318e08394c5f\") " pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:31:57 crc kubenswrapper[4771]: I1002 11:31:57.175429 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2hhg\" (UniqueName: \"kubernetes.io/projected/d5c9d376-d940-4f03-84b0-318e08394c5f-kube-api-access-k2hhg\") pod \"certified-operators-xlvnh\" (UID: \"d5c9d376-d940-4f03-84b0-318e08394c5f\") " pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:31:57 crc kubenswrapper[4771]: I1002 11:31:57.175484 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c9d376-d940-4f03-84b0-318e08394c5f-utilities\") pod \"certified-operators-xlvnh\" (UID: \"d5c9d376-d940-4f03-84b0-318e08394c5f\") " pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:31:57 crc kubenswrapper[4771]: I1002 11:31:57.175977 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c9d376-d940-4f03-84b0-318e08394c5f-catalog-content\") pod \"certified-operators-xlvnh\" (UID: \"d5c9d376-d940-4f03-84b0-318e08394c5f\") " pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:31:57 crc kubenswrapper[4771]: I1002 11:31:57.176023 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c9d376-d940-4f03-84b0-318e08394c5f-utilities\") pod \"certified-operators-xlvnh\" (UID: \"d5c9d376-d940-4f03-84b0-318e08394c5f\") " pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:31:57 crc kubenswrapper[4771]: I1002 11:31:57.208572 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2hhg\" (UniqueName: \"kubernetes.io/projected/d5c9d376-d940-4f03-84b0-318e08394c5f-kube-api-access-k2hhg\") pod \"certified-operators-xlvnh\" (UID: \"d5c9d376-d940-4f03-84b0-318e08394c5f\") " pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:31:57 crc kubenswrapper[4771]: I1002 11:31:57.298023 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:31:57 crc kubenswrapper[4771]: I1002 11:31:57.887154 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xlvnh"] Oct 02 11:31:58 crc kubenswrapper[4771]: I1002 11:31:58.005721 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlvnh" event={"ID":"d5c9d376-d940-4f03-84b0-318e08394c5f","Type":"ContainerStarted","Data":"55f09c9b295cb3e021714e3e706ed0f513fd8ba3426ca24ca7ad1bc5f4092224"} Oct 02 11:31:59 crc kubenswrapper[4771]: I1002 11:31:59.018790 4771 generic.go:334] "Generic (PLEG): container finished" podID="d5c9d376-d940-4f03-84b0-318e08394c5f" containerID="d7a866ae2f65e21389ea5e9018e9e6931c90d3e893268bee302e5cd9e0b0d70b" exitCode=0 Oct 02 11:31:59 crc kubenswrapper[4771]: I1002 11:31:59.018850 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlvnh" event={"ID":"d5c9d376-d940-4f03-84b0-318e08394c5f","Type":"ContainerDied","Data":"d7a866ae2f65e21389ea5e9018e9e6931c90d3e893268bee302e5cd9e0b0d70b"} Oct 02 11:32:01 crc kubenswrapper[4771]: I1002 11:32:01.043789 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlvnh" event={"ID":"d5c9d376-d940-4f03-84b0-318e08394c5f","Type":"ContainerStarted","Data":"a95ae9a1a1a4526c53f0a214967cace838c073d1909448270027bea58f92a8aa"} Oct 02 11:32:02 crc kubenswrapper[4771]: E1002 11:32:02.641029 4771 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5c9d376_d940_4f03_84b0_318e08394c5f.slice/crio-a95ae9a1a1a4526c53f0a214967cace838c073d1909448270027bea58f92a8aa.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5c9d376_d940_4f03_84b0_318e08394c5f.slice/crio-conmon-a95ae9a1a1a4526c53f0a214967cace838c073d1909448270027bea58f92a8aa.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:32:02 crc kubenswrapper[4771]: E1002 11:32:02.925375 4771 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.53:36842->38.102.83.53:41045: write tcp 38.102.83.53:36842->38.102.83.53:41045: write: connection reset by peer Oct 02 11:32:03 crc kubenswrapper[4771]: I1002 11:32:03.070179 4771 generic.go:334] "Generic (PLEG): container finished" podID="d5c9d376-d940-4f03-84b0-318e08394c5f" containerID="a95ae9a1a1a4526c53f0a214967cace838c073d1909448270027bea58f92a8aa" exitCode=0 Oct 02 11:32:03 crc kubenswrapper[4771]: I1002 11:32:03.070232 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlvnh" event={"ID":"d5c9d376-d940-4f03-84b0-318e08394c5f","Type":"ContainerDied","Data":"a95ae9a1a1a4526c53f0a214967cace838c073d1909448270027bea58f92a8aa"} Oct 02 11:32:04 crc kubenswrapper[4771]: I1002 11:32:04.086575 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlvnh" event={"ID":"d5c9d376-d940-4f03-84b0-318e08394c5f","Type":"ContainerStarted","Data":"d2e4be0ba2bbbc511e8143d26b7540cf7ded6b3aa3b697ab17595359bea52a2d"} Oct 02 11:32:04 crc kubenswrapper[4771]: I1002 11:32:04.126698 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xlvnh" podStartSLOduration=3.565213147 podStartE2EDuration="8.12666674s" podCreationTimestamp="2025-10-02 11:31:56 +0000 UTC" firstStartedPulling="2025-10-02 11:31:59.021462308 +0000 UTC m=+6906.669147365" lastFinishedPulling="2025-10-02 11:32:03.582915891 +0000 UTC m=+6911.230600958" observedRunningTime="2025-10-02 11:32:04.109462927 +0000 UTC m=+6911.757148014" watchObservedRunningTime="2025-10-02 11:32:04.12666674 +0000 UTC m=+6911.774351807" Oct 02 11:32:07 crc kubenswrapper[4771]: I1002 11:32:07.298422 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:32:07 crc kubenswrapper[4771]: I1002 11:32:07.298735 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:32:07 crc kubenswrapper[4771]: I1002 11:32:07.353687 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:32:08 crc kubenswrapper[4771]: I1002 11:32:08.188924 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:32:08 crc kubenswrapper[4771]: I1002 11:32:08.245373 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xlvnh"] Oct 02 11:32:10 crc kubenswrapper[4771]: I1002 11:32:10.154458 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xlvnh" podUID="d5c9d376-d940-4f03-84b0-318e08394c5f" containerName="registry-server" containerID="cri-o://d2e4be0ba2bbbc511e8143d26b7540cf7ded6b3aa3b697ab17595359bea52a2d" gracePeriod=2 Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.183501 4771 generic.go:334] "Generic (PLEG): container finished" podID="d5c9d376-d940-4f03-84b0-318e08394c5f" containerID="d2e4be0ba2bbbc511e8143d26b7540cf7ded6b3aa3b697ab17595359bea52a2d" exitCode=0 Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.183665 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlvnh" event={"ID":"d5c9d376-d940-4f03-84b0-318e08394c5f","Type":"ContainerDied","Data":"d2e4be0ba2bbbc511e8143d26b7540cf7ded6b3aa3b697ab17595359bea52a2d"} Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.184000 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlvnh" event={"ID":"d5c9d376-d940-4f03-84b0-318e08394c5f","Type":"ContainerDied","Data":"55f09c9b295cb3e021714e3e706ed0f513fd8ba3426ca24ca7ad1bc5f4092224"} Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.184026 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55f09c9b295cb3e021714e3e706ed0f513fd8ba3426ca24ca7ad1bc5f4092224" Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.225704 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.357654 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c9d376-d940-4f03-84b0-318e08394c5f-catalog-content\") pod \"d5c9d376-d940-4f03-84b0-318e08394c5f\" (UID: \"d5c9d376-d940-4f03-84b0-318e08394c5f\") " Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.357880 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2hhg\" (UniqueName: \"kubernetes.io/projected/d5c9d376-d940-4f03-84b0-318e08394c5f-kube-api-access-k2hhg\") pod \"d5c9d376-d940-4f03-84b0-318e08394c5f\" (UID: \"d5c9d376-d940-4f03-84b0-318e08394c5f\") " Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.358158 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c9d376-d940-4f03-84b0-318e08394c5f-utilities\") pod \"d5c9d376-d940-4f03-84b0-318e08394c5f\" (UID: \"d5c9d376-d940-4f03-84b0-318e08394c5f\") " Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.359192 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5c9d376-d940-4f03-84b0-318e08394c5f-utilities" (OuterVolumeSpecName: "utilities") pod "d5c9d376-d940-4f03-84b0-318e08394c5f" (UID: "d5c9d376-d940-4f03-84b0-318e08394c5f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.380693 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5c9d376-d940-4f03-84b0-318e08394c5f-kube-api-access-k2hhg" (OuterVolumeSpecName: "kube-api-access-k2hhg") pod "d5c9d376-d940-4f03-84b0-318e08394c5f" (UID: "d5c9d376-d940-4f03-84b0-318e08394c5f"). InnerVolumeSpecName "kube-api-access-k2hhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.413316 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5c9d376-d940-4f03-84b0-318e08394c5f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5c9d376-d940-4f03-84b0-318e08394c5f" (UID: "d5c9d376-d940-4f03-84b0-318e08394c5f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.461424 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c9d376-d940-4f03-84b0-318e08394c5f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.461470 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c9d376-d940-4f03-84b0-318e08394c5f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:11 crc kubenswrapper[4771]: I1002 11:32:11.461483 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2hhg\" (UniqueName: \"kubernetes.io/projected/d5c9d376-d940-4f03-84b0-318e08394c5f-kube-api-access-k2hhg\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:12 crc kubenswrapper[4771]: I1002 11:32:12.146122 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:32:12 crc kubenswrapper[4771]: I1002 11:32:12.146510 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:32:12 crc kubenswrapper[4771]: I1002 11:32:12.146556 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 11:32:12 crc kubenswrapper[4771]: I1002 11:32:12.147437 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2486e5a32ac351379d45c0c9b8440f7fe86fe9b348bd28df0df4a5133b9cb5b8"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:32:12 crc kubenswrapper[4771]: I1002 11:32:12.147498 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://2486e5a32ac351379d45c0c9b8440f7fe86fe9b348bd28df0df4a5133b9cb5b8" gracePeriod=600 Oct 02 11:32:12 crc kubenswrapper[4771]: I1002 11:32:12.197042 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlvnh" Oct 02 11:32:12 crc kubenswrapper[4771]: I1002 11:32:12.221522 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xlvnh"] Oct 02 11:32:12 crc kubenswrapper[4771]: I1002 11:32:12.233874 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xlvnh"] Oct 02 11:32:13 crc kubenswrapper[4771]: I1002 11:32:13.213827 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="2486e5a32ac351379d45c0c9b8440f7fe86fe9b348bd28df0df4a5133b9cb5b8" exitCode=0 Oct 02 11:32:13 crc kubenswrapper[4771]: I1002 11:32:13.213911 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"2486e5a32ac351379d45c0c9b8440f7fe86fe9b348bd28df0df4a5133b9cb5b8"} Oct 02 11:32:13 crc kubenswrapper[4771]: I1002 11:32:13.214624 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9"} Oct 02 11:32:13 crc kubenswrapper[4771]: I1002 11:32:13.214654 4771 scope.go:117] "RemoveContainer" containerID="69884cee3147175d39dac69d4e9ce76a509bd150904c7df9740e1cdb81c5eecc" Oct 02 11:32:13 crc kubenswrapper[4771]: I1002 11:32:13.708191 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5c9d376-d940-4f03-84b0-318e08394c5f" path="/var/lib/kubelet/pods/d5c9d376-d940-4f03-84b0-318e08394c5f/volumes" Oct 02 11:33:08 crc kubenswrapper[4771]: I1002 11:33:08.982080 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5mzp7"] Oct 02 11:33:08 crc kubenswrapper[4771]: E1002 11:33:08.983552 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c9d376-d940-4f03-84b0-318e08394c5f" containerName="extract-content" Oct 02 11:33:08 crc kubenswrapper[4771]: I1002 11:33:08.983573 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c9d376-d940-4f03-84b0-318e08394c5f" containerName="extract-content" Oct 02 11:33:08 crc kubenswrapper[4771]: E1002 11:33:08.983619 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c9d376-d940-4f03-84b0-318e08394c5f" containerName="extract-utilities" Oct 02 11:33:08 crc kubenswrapper[4771]: I1002 11:33:08.983630 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c9d376-d940-4f03-84b0-318e08394c5f" containerName="extract-utilities" Oct 02 11:33:08 crc kubenswrapper[4771]: E1002 11:33:08.983647 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c9d376-d940-4f03-84b0-318e08394c5f" containerName="registry-server" Oct 02 11:33:08 crc kubenswrapper[4771]: I1002 11:33:08.983655 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c9d376-d940-4f03-84b0-318e08394c5f" containerName="registry-server" Oct 02 11:33:08 crc kubenswrapper[4771]: I1002 11:33:08.984019 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5c9d376-d940-4f03-84b0-318e08394c5f" containerName="registry-server" Oct 02 11:33:09 crc kubenswrapper[4771]: I1002 11:33:09.004858 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5mzp7"] Oct 02 11:33:09 crc kubenswrapper[4771]: I1002 11:33:09.019757 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:09 crc kubenswrapper[4771]: I1002 11:33:09.105657 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdqwm\" (UniqueName: \"kubernetes.io/projected/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-kube-api-access-vdqwm\") pod \"redhat-marketplace-5mzp7\" (UID: \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\") " pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:09 crc kubenswrapper[4771]: I1002 11:33:09.106293 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-catalog-content\") pod \"redhat-marketplace-5mzp7\" (UID: \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\") " pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:09 crc kubenswrapper[4771]: I1002 11:33:09.106515 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-utilities\") pod \"redhat-marketplace-5mzp7\" (UID: \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\") " pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:09 crc kubenswrapper[4771]: I1002 11:33:09.209437 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-catalog-content\") pod \"redhat-marketplace-5mzp7\" (UID: \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\") " pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:09 crc kubenswrapper[4771]: I1002 11:33:09.209503 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-utilities\") pod \"redhat-marketplace-5mzp7\" (UID: \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\") " pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:09 crc kubenswrapper[4771]: I1002 11:33:09.209565 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdqwm\" (UniqueName: \"kubernetes.io/projected/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-kube-api-access-vdqwm\") pod \"redhat-marketplace-5mzp7\" (UID: \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\") " pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:09 crc kubenswrapper[4771]: I1002 11:33:09.210006 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-catalog-content\") pod \"redhat-marketplace-5mzp7\" (UID: \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\") " pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:09 crc kubenswrapper[4771]: I1002 11:33:09.210050 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-utilities\") pod \"redhat-marketplace-5mzp7\" (UID: \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\") " pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:09 crc kubenswrapper[4771]: I1002 11:33:09.231487 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdqwm\" (UniqueName: \"kubernetes.io/projected/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-kube-api-access-vdqwm\") pod \"redhat-marketplace-5mzp7\" (UID: \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\") " pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:09 crc kubenswrapper[4771]: I1002 11:33:09.358098 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:09 crc kubenswrapper[4771]: I1002 11:33:09.853948 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5mzp7"] Oct 02 11:33:10 crc kubenswrapper[4771]: I1002 11:33:10.859753 4771 generic.go:334] "Generic (PLEG): container finished" podID="bcee95b3-b4ce-4839-b6b9-7c8ca171550a" containerID="3e6cc3ac4fb0acd0a8a374e95f1b7872e295dd19b91a6e5e4c821c480f2c04fc" exitCode=0 Oct 02 11:33:10 crc kubenswrapper[4771]: I1002 11:33:10.859809 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mzp7" event={"ID":"bcee95b3-b4ce-4839-b6b9-7c8ca171550a","Type":"ContainerDied","Data":"3e6cc3ac4fb0acd0a8a374e95f1b7872e295dd19b91a6e5e4c821c480f2c04fc"} Oct 02 11:33:10 crc kubenswrapper[4771]: I1002 11:33:10.860070 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mzp7" event={"ID":"bcee95b3-b4ce-4839-b6b9-7c8ca171550a","Type":"ContainerStarted","Data":"7c1a4bf4bd32e90ec10296a9cf6dbade202f0fe9c105173e2787dad3938f35c1"} Oct 02 11:33:10 crc kubenswrapper[4771]: I1002 11:33:10.862190 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:33:11 crc kubenswrapper[4771]: I1002 11:33:11.874105 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mzp7" event={"ID":"bcee95b3-b4ce-4839-b6b9-7c8ca171550a","Type":"ContainerStarted","Data":"3f23de423b82947983c0c7edc5951a6073d5909b3b0f5d7ade8b5437c9fc6dbc"} Oct 02 11:33:12 crc kubenswrapper[4771]: I1002 11:33:12.884365 4771 generic.go:334] "Generic (PLEG): container finished" podID="bcee95b3-b4ce-4839-b6b9-7c8ca171550a" containerID="3f23de423b82947983c0c7edc5951a6073d5909b3b0f5d7ade8b5437c9fc6dbc" exitCode=0 Oct 02 11:33:12 crc kubenswrapper[4771]: I1002 11:33:12.884404 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mzp7" event={"ID":"bcee95b3-b4ce-4839-b6b9-7c8ca171550a","Type":"ContainerDied","Data":"3f23de423b82947983c0c7edc5951a6073d5909b3b0f5d7ade8b5437c9fc6dbc"} Oct 02 11:33:13 crc kubenswrapper[4771]: I1002 11:33:13.898169 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mzp7" event={"ID":"bcee95b3-b4ce-4839-b6b9-7c8ca171550a","Type":"ContainerStarted","Data":"de2cd5a6ef08f0551e723cc3ecc15e0de35bcf643f541a82823f8db6743b1380"} Oct 02 11:33:19 crc kubenswrapper[4771]: I1002 11:33:19.358911 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:19 crc kubenswrapper[4771]: I1002 11:33:19.359495 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:19 crc kubenswrapper[4771]: I1002 11:33:19.411821 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:19 crc kubenswrapper[4771]: I1002 11:33:19.432016 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5mzp7" podStartSLOduration=8.848412353 podStartE2EDuration="11.431997673s" podCreationTimestamp="2025-10-02 11:33:08 +0000 UTC" firstStartedPulling="2025-10-02 11:33:10.861943291 +0000 UTC m=+6978.509628358" lastFinishedPulling="2025-10-02 11:33:13.445528611 +0000 UTC m=+6981.093213678" observedRunningTime="2025-10-02 11:33:13.916755749 +0000 UTC m=+6981.564440816" watchObservedRunningTime="2025-10-02 11:33:19.431997673 +0000 UTC m=+6987.079682740" Oct 02 11:33:20 crc kubenswrapper[4771]: I1002 11:33:20.057183 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:20 crc kubenswrapper[4771]: I1002 11:33:20.117591 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5mzp7"] Oct 02 11:33:22 crc kubenswrapper[4771]: I1002 11:33:22.013042 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5mzp7" podUID="bcee95b3-b4ce-4839-b6b9-7c8ca171550a" containerName="registry-server" containerID="cri-o://de2cd5a6ef08f0551e723cc3ecc15e0de35bcf643f541a82823f8db6743b1380" gracePeriod=2 Oct 02 11:33:22 crc kubenswrapper[4771]: I1002 11:33:22.585892 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:22 crc kubenswrapper[4771]: I1002 11:33:22.779424 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-utilities\") pod \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\" (UID: \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\") " Oct 02 11:33:22 crc kubenswrapper[4771]: I1002 11:33:22.779556 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-catalog-content\") pod \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\" (UID: \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\") " Oct 02 11:33:22 crc kubenswrapper[4771]: I1002 11:33:22.779828 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdqwm\" (UniqueName: \"kubernetes.io/projected/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-kube-api-access-vdqwm\") pod \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\" (UID: \"bcee95b3-b4ce-4839-b6b9-7c8ca171550a\") " Oct 02 11:33:22 crc kubenswrapper[4771]: I1002 11:33:22.783297 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-utilities" (OuterVolumeSpecName: "utilities") pod "bcee95b3-b4ce-4839-b6b9-7c8ca171550a" (UID: "bcee95b3-b4ce-4839-b6b9-7c8ca171550a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:33:22 crc kubenswrapper[4771]: I1002 11:33:22.791044 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-kube-api-access-vdqwm" (OuterVolumeSpecName: "kube-api-access-vdqwm") pod "bcee95b3-b4ce-4839-b6b9-7c8ca171550a" (UID: "bcee95b3-b4ce-4839-b6b9-7c8ca171550a"). InnerVolumeSpecName "kube-api-access-vdqwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:33:22 crc kubenswrapper[4771]: I1002 11:33:22.804673 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bcee95b3-b4ce-4839-b6b9-7c8ca171550a" (UID: "bcee95b3-b4ce-4839-b6b9-7c8ca171550a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:33:22 crc kubenswrapper[4771]: I1002 11:33:22.884644 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:33:22 crc kubenswrapper[4771]: I1002 11:33:22.884697 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:33:22 crc kubenswrapper[4771]: I1002 11:33:22.884713 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdqwm\" (UniqueName: \"kubernetes.io/projected/bcee95b3-b4ce-4839-b6b9-7c8ca171550a-kube-api-access-vdqwm\") on node \"crc\" DevicePath \"\"" Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.038518 4771 generic.go:334] "Generic (PLEG): container finished" podID="bcee95b3-b4ce-4839-b6b9-7c8ca171550a" containerID="de2cd5a6ef08f0551e723cc3ecc15e0de35bcf643f541a82823f8db6743b1380" exitCode=0 Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.038589 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mzp7" event={"ID":"bcee95b3-b4ce-4839-b6b9-7c8ca171550a","Type":"ContainerDied","Data":"de2cd5a6ef08f0551e723cc3ecc15e0de35bcf643f541a82823f8db6743b1380"} Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.038630 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mzp7" event={"ID":"bcee95b3-b4ce-4839-b6b9-7c8ca171550a","Type":"ContainerDied","Data":"7c1a4bf4bd32e90ec10296a9cf6dbade202f0fe9c105173e2787dad3938f35c1"} Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.038622 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5mzp7" Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.038652 4771 scope.go:117] "RemoveContainer" containerID="de2cd5a6ef08f0551e723cc3ecc15e0de35bcf643f541a82823f8db6743b1380" Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.074497 4771 scope.go:117] "RemoveContainer" containerID="3f23de423b82947983c0c7edc5951a6073d5909b3b0f5d7ade8b5437c9fc6dbc" Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.088067 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5mzp7"] Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.104536 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5mzp7"] Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.107910 4771 scope.go:117] "RemoveContainer" containerID="3e6cc3ac4fb0acd0a8a374e95f1b7872e295dd19b91a6e5e4c821c480f2c04fc" Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.186673 4771 scope.go:117] "RemoveContainer" containerID="de2cd5a6ef08f0551e723cc3ecc15e0de35bcf643f541a82823f8db6743b1380" Oct 02 11:33:23 crc kubenswrapper[4771]: E1002 11:33:23.188216 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de2cd5a6ef08f0551e723cc3ecc15e0de35bcf643f541a82823f8db6743b1380\": container with ID starting with de2cd5a6ef08f0551e723cc3ecc15e0de35bcf643f541a82823f8db6743b1380 not found: ID does not exist" containerID="de2cd5a6ef08f0551e723cc3ecc15e0de35bcf643f541a82823f8db6743b1380" Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.188276 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de2cd5a6ef08f0551e723cc3ecc15e0de35bcf643f541a82823f8db6743b1380"} err="failed to get container status \"de2cd5a6ef08f0551e723cc3ecc15e0de35bcf643f541a82823f8db6743b1380\": rpc error: code = NotFound desc = could not find container \"de2cd5a6ef08f0551e723cc3ecc15e0de35bcf643f541a82823f8db6743b1380\": container with ID starting with de2cd5a6ef08f0551e723cc3ecc15e0de35bcf643f541a82823f8db6743b1380 not found: ID does not exist" Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.188318 4771 scope.go:117] "RemoveContainer" containerID="3f23de423b82947983c0c7edc5951a6073d5909b3b0f5d7ade8b5437c9fc6dbc" Oct 02 11:33:23 crc kubenswrapper[4771]: E1002 11:33:23.188952 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f23de423b82947983c0c7edc5951a6073d5909b3b0f5d7ade8b5437c9fc6dbc\": container with ID starting with 3f23de423b82947983c0c7edc5951a6073d5909b3b0f5d7ade8b5437c9fc6dbc not found: ID does not exist" containerID="3f23de423b82947983c0c7edc5951a6073d5909b3b0f5d7ade8b5437c9fc6dbc" Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.189018 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f23de423b82947983c0c7edc5951a6073d5909b3b0f5d7ade8b5437c9fc6dbc"} err="failed to get container status \"3f23de423b82947983c0c7edc5951a6073d5909b3b0f5d7ade8b5437c9fc6dbc\": rpc error: code = NotFound desc = could not find container \"3f23de423b82947983c0c7edc5951a6073d5909b3b0f5d7ade8b5437c9fc6dbc\": container with ID starting with 3f23de423b82947983c0c7edc5951a6073d5909b3b0f5d7ade8b5437c9fc6dbc not found: ID does not exist" Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.189041 4771 scope.go:117] "RemoveContainer" containerID="3e6cc3ac4fb0acd0a8a374e95f1b7872e295dd19b91a6e5e4c821c480f2c04fc" Oct 02 11:33:23 crc kubenswrapper[4771]: E1002 11:33:23.189527 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e6cc3ac4fb0acd0a8a374e95f1b7872e295dd19b91a6e5e4c821c480f2c04fc\": container with ID starting with 3e6cc3ac4fb0acd0a8a374e95f1b7872e295dd19b91a6e5e4c821c480f2c04fc not found: ID does not exist" containerID="3e6cc3ac4fb0acd0a8a374e95f1b7872e295dd19b91a6e5e4c821c480f2c04fc" Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.189604 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e6cc3ac4fb0acd0a8a374e95f1b7872e295dd19b91a6e5e4c821c480f2c04fc"} err="failed to get container status \"3e6cc3ac4fb0acd0a8a374e95f1b7872e295dd19b91a6e5e4c821c480f2c04fc\": rpc error: code = NotFound desc = could not find container \"3e6cc3ac4fb0acd0a8a374e95f1b7872e295dd19b91a6e5e4c821c480f2c04fc\": container with ID starting with 3e6cc3ac4fb0acd0a8a374e95f1b7872e295dd19b91a6e5e4c821c480f2c04fc not found: ID does not exist" Oct 02 11:33:23 crc kubenswrapper[4771]: I1002 11:33:23.699285 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcee95b3-b4ce-4839-b6b9-7c8ca171550a" path="/var/lib/kubelet/pods/bcee95b3-b4ce-4839-b6b9-7c8ca171550a/volumes" Oct 02 11:34:12 crc kubenswrapper[4771]: I1002 11:34:12.146240 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:34:12 crc kubenswrapper[4771]: I1002 11:34:12.146932 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:34:42 crc kubenswrapper[4771]: I1002 11:34:42.145934 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:34:42 crc kubenswrapper[4771]: I1002 11:34:42.146646 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:35:12 crc kubenswrapper[4771]: I1002 11:35:12.146215 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:35:12 crc kubenswrapper[4771]: I1002 11:35:12.147102 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:35:12 crc kubenswrapper[4771]: I1002 11:35:12.147208 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 11:35:12 crc kubenswrapper[4771]: I1002 11:35:12.148776 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:35:12 crc kubenswrapper[4771]: I1002 11:35:12.148844 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" gracePeriod=600 Oct 02 11:35:12 crc kubenswrapper[4771]: E1002 11:35:12.282604 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:35:12 crc kubenswrapper[4771]: I1002 11:35:12.397808 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" exitCode=0 Oct 02 11:35:12 crc kubenswrapper[4771]: I1002 11:35:12.397895 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9"} Oct 02 11:35:12 crc kubenswrapper[4771]: I1002 11:35:12.397982 4771 scope.go:117] "RemoveContainer" containerID="2486e5a32ac351379d45c0c9b8440f7fe86fe9b348bd28df0df4a5133b9cb5b8" Oct 02 11:35:12 crc kubenswrapper[4771]: I1002 11:35:12.399403 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:35:12 crc kubenswrapper[4771]: E1002 11:35:12.399882 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:35:24 crc kubenswrapper[4771]: I1002 11:35:24.682388 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:35:24 crc kubenswrapper[4771]: E1002 11:35:24.683604 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:35:38 crc kubenswrapper[4771]: I1002 11:35:38.683457 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:35:38 crc kubenswrapper[4771]: E1002 11:35:38.684485 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:35:53 crc kubenswrapper[4771]: I1002 11:35:53.690219 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:35:53 crc kubenswrapper[4771]: E1002 11:35:53.692573 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:36:04 crc kubenswrapper[4771]: I1002 11:36:04.681899 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:36:04 crc kubenswrapper[4771]: E1002 11:36:04.683039 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:36:17 crc kubenswrapper[4771]: I1002 11:36:17.681157 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:36:17 crc kubenswrapper[4771]: E1002 11:36:17.682026 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.405815 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7xmmr"] Oct 02 11:36:22 crc kubenswrapper[4771]: E1002 11:36:22.407015 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcee95b3-b4ce-4839-b6b9-7c8ca171550a" containerName="extract-content" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.407036 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcee95b3-b4ce-4839-b6b9-7c8ca171550a" containerName="extract-content" Oct 02 11:36:22 crc kubenswrapper[4771]: E1002 11:36:22.407074 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcee95b3-b4ce-4839-b6b9-7c8ca171550a" containerName="registry-server" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.407084 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcee95b3-b4ce-4839-b6b9-7c8ca171550a" containerName="registry-server" Oct 02 11:36:22 crc kubenswrapper[4771]: E1002 11:36:22.407100 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcee95b3-b4ce-4839-b6b9-7c8ca171550a" containerName="extract-utilities" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.407109 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcee95b3-b4ce-4839-b6b9-7c8ca171550a" containerName="extract-utilities" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.407461 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcee95b3-b4ce-4839-b6b9-7c8ca171550a" containerName="registry-server" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.409727 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.440111 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7xmmr"] Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.514020 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvq2c\" (UniqueName: \"kubernetes.io/projected/a70441f1-e293-4c49-8206-1c1806927941-kube-api-access-wvq2c\") pod \"redhat-operators-7xmmr\" (UID: \"a70441f1-e293-4c49-8206-1c1806927941\") " pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.514193 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a70441f1-e293-4c49-8206-1c1806927941-catalog-content\") pod \"redhat-operators-7xmmr\" (UID: \"a70441f1-e293-4c49-8206-1c1806927941\") " pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.514446 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a70441f1-e293-4c49-8206-1c1806927941-utilities\") pod \"redhat-operators-7xmmr\" (UID: \"a70441f1-e293-4c49-8206-1c1806927941\") " pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.617453 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a70441f1-e293-4c49-8206-1c1806927941-catalog-content\") pod \"redhat-operators-7xmmr\" (UID: \"a70441f1-e293-4c49-8206-1c1806927941\") " pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.617706 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a70441f1-e293-4c49-8206-1c1806927941-utilities\") pod \"redhat-operators-7xmmr\" (UID: \"a70441f1-e293-4c49-8206-1c1806927941\") " pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.617803 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvq2c\" (UniqueName: \"kubernetes.io/projected/a70441f1-e293-4c49-8206-1c1806927941-kube-api-access-wvq2c\") pod \"redhat-operators-7xmmr\" (UID: \"a70441f1-e293-4c49-8206-1c1806927941\") " pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.617996 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a70441f1-e293-4c49-8206-1c1806927941-catalog-content\") pod \"redhat-operators-7xmmr\" (UID: \"a70441f1-e293-4c49-8206-1c1806927941\") " pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.618357 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a70441f1-e293-4c49-8206-1c1806927941-utilities\") pod \"redhat-operators-7xmmr\" (UID: \"a70441f1-e293-4c49-8206-1c1806927941\") " pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.641259 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvq2c\" (UniqueName: \"kubernetes.io/projected/a70441f1-e293-4c49-8206-1c1806927941-kube-api-access-wvq2c\") pod \"redhat-operators-7xmmr\" (UID: \"a70441f1-e293-4c49-8206-1c1806927941\") " pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:36:22 crc kubenswrapper[4771]: I1002 11:36:22.748137 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:36:23 crc kubenswrapper[4771]: I1002 11:36:23.453551 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7xmmr"] Oct 02 11:36:24 crc kubenswrapper[4771]: I1002 11:36:24.300626 4771 generic.go:334] "Generic (PLEG): container finished" podID="a70441f1-e293-4c49-8206-1c1806927941" containerID="c5a329e0e5cf44e57bffbf6d9a5c1b911ffedcbcaf5d241195fe4b896e864099" exitCode=0 Oct 02 11:36:24 crc kubenswrapper[4771]: I1002 11:36:24.300695 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xmmr" event={"ID":"a70441f1-e293-4c49-8206-1c1806927941","Type":"ContainerDied","Data":"c5a329e0e5cf44e57bffbf6d9a5c1b911ffedcbcaf5d241195fe4b896e864099"} Oct 02 11:36:24 crc kubenswrapper[4771]: I1002 11:36:24.300946 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xmmr" event={"ID":"a70441f1-e293-4c49-8206-1c1806927941","Type":"ContainerStarted","Data":"79820260815b6f91c06257c3369b5d3c5d35c586160a09b2ae9364f3543c4534"} Oct 02 11:36:26 crc kubenswrapper[4771]: I1002 11:36:26.328435 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xmmr" event={"ID":"a70441f1-e293-4c49-8206-1c1806927941","Type":"ContainerStarted","Data":"8f2f83868c1a1f8052fc7665e25e6c04eb62225f5b33b6b064df7df2fda81daf"} Oct 02 11:36:29 crc kubenswrapper[4771]: I1002 11:36:29.376383 4771 generic.go:334] "Generic (PLEG): container finished" podID="a70441f1-e293-4c49-8206-1c1806927941" containerID="8f2f83868c1a1f8052fc7665e25e6c04eb62225f5b33b6b064df7df2fda81daf" exitCode=0 Oct 02 11:36:29 crc kubenswrapper[4771]: I1002 11:36:29.376457 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xmmr" event={"ID":"a70441f1-e293-4c49-8206-1c1806927941","Type":"ContainerDied","Data":"8f2f83868c1a1f8052fc7665e25e6c04eb62225f5b33b6b064df7df2fda81daf"} Oct 02 11:36:29 crc kubenswrapper[4771]: I1002 11:36:29.682063 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:36:29 crc kubenswrapper[4771]: E1002 11:36:29.682455 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:36:30 crc kubenswrapper[4771]: I1002 11:36:30.390843 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xmmr" event={"ID":"a70441f1-e293-4c49-8206-1c1806927941","Type":"ContainerStarted","Data":"affa16a74eb0604b86c3314a14ed76c9737a8b909faa51015aa11f11a5e5ed87"} Oct 02 11:36:30 crc kubenswrapper[4771]: I1002 11:36:30.418580 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7xmmr" podStartSLOduration=2.84471325 podStartE2EDuration="8.418551454s" podCreationTimestamp="2025-10-02 11:36:22 +0000 UTC" firstStartedPulling="2025-10-02 11:36:24.30327235 +0000 UTC m=+7171.950957417" lastFinishedPulling="2025-10-02 11:36:29.877110554 +0000 UTC m=+7177.524795621" observedRunningTime="2025-10-02 11:36:30.412240421 +0000 UTC m=+7178.059925488" watchObservedRunningTime="2025-10-02 11:36:30.418551454 +0000 UTC m=+7178.066236521" Oct 02 11:36:32 crc kubenswrapper[4771]: I1002 11:36:32.748631 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:36:32 crc kubenswrapper[4771]: I1002 11:36:32.749195 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:36:33 crc kubenswrapper[4771]: I1002 11:36:33.803393 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7xmmr" podUID="a70441f1-e293-4c49-8206-1c1806927941" containerName="registry-server" probeResult="failure" output=< Oct 02 11:36:33 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:36:33 crc kubenswrapper[4771]: > Oct 02 11:36:42 crc kubenswrapper[4771]: I1002 11:36:42.682990 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:36:42 crc kubenswrapper[4771]: E1002 11:36:42.684310 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:36:43 crc kubenswrapper[4771]: I1002 11:36:43.800417 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7xmmr" podUID="a70441f1-e293-4c49-8206-1c1806927941" containerName="registry-server" probeResult="failure" output=< Oct 02 11:36:43 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:36:43 crc kubenswrapper[4771]: > Oct 02 11:36:53 crc kubenswrapper[4771]: I1002 11:36:53.797346 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7xmmr" podUID="a70441f1-e293-4c49-8206-1c1806927941" containerName="registry-server" probeResult="failure" output=< Oct 02 11:36:53 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:36:53 crc kubenswrapper[4771]: > Oct 02 11:36:54 crc kubenswrapper[4771]: I1002 11:36:54.681791 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:36:54 crc kubenswrapper[4771]: E1002 11:36:54.682489 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:37:02 crc kubenswrapper[4771]: I1002 11:37:02.796552 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:37:02 crc kubenswrapper[4771]: I1002 11:37:02.868601 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:37:03 crc kubenswrapper[4771]: I1002 11:37:03.051334 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7xmmr"] Oct 02 11:37:03 crc kubenswrapper[4771]: I1002 11:37:03.856098 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7xmmr" podUID="a70441f1-e293-4c49-8206-1c1806927941" containerName="registry-server" containerID="cri-o://affa16a74eb0604b86c3314a14ed76c9737a8b909faa51015aa11f11a5e5ed87" gracePeriod=2 Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.503510 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.628791 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a70441f1-e293-4c49-8206-1c1806927941-utilities\") pod \"a70441f1-e293-4c49-8206-1c1806927941\" (UID: \"a70441f1-e293-4c49-8206-1c1806927941\") " Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.628860 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a70441f1-e293-4c49-8206-1c1806927941-catalog-content\") pod \"a70441f1-e293-4c49-8206-1c1806927941\" (UID: \"a70441f1-e293-4c49-8206-1c1806927941\") " Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.629122 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvq2c\" (UniqueName: \"kubernetes.io/projected/a70441f1-e293-4c49-8206-1c1806927941-kube-api-access-wvq2c\") pod \"a70441f1-e293-4c49-8206-1c1806927941\" (UID: \"a70441f1-e293-4c49-8206-1c1806927941\") " Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.629853 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a70441f1-e293-4c49-8206-1c1806927941-utilities" (OuterVolumeSpecName: "utilities") pod "a70441f1-e293-4c49-8206-1c1806927941" (UID: "a70441f1-e293-4c49-8206-1c1806927941"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.634649 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a70441f1-e293-4c49-8206-1c1806927941-kube-api-access-wvq2c" (OuterVolumeSpecName: "kube-api-access-wvq2c") pod "a70441f1-e293-4c49-8206-1c1806927941" (UID: "a70441f1-e293-4c49-8206-1c1806927941"). InnerVolumeSpecName "kube-api-access-wvq2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.731987 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a70441f1-e293-4c49-8206-1c1806927941-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a70441f1-e293-4c49-8206-1c1806927941" (UID: "a70441f1-e293-4c49-8206-1c1806927941"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.733155 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a70441f1-e293-4c49-8206-1c1806927941-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.733189 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a70441f1-e293-4c49-8206-1c1806927941-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.733201 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvq2c\" (UniqueName: \"kubernetes.io/projected/a70441f1-e293-4c49-8206-1c1806927941-kube-api-access-wvq2c\") on node \"crc\" DevicePath \"\"" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.870696 4771 generic.go:334] "Generic (PLEG): container finished" podID="a70441f1-e293-4c49-8206-1c1806927941" containerID="affa16a74eb0604b86c3314a14ed76c9737a8b909faa51015aa11f11a5e5ed87" exitCode=0 Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.870746 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xmmr" event={"ID":"a70441f1-e293-4c49-8206-1c1806927941","Type":"ContainerDied","Data":"affa16a74eb0604b86c3314a14ed76c9737a8b909faa51015aa11f11a5e5ed87"} Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.870786 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xmmr" event={"ID":"a70441f1-e293-4c49-8206-1c1806927941","Type":"ContainerDied","Data":"79820260815b6f91c06257c3369b5d3c5d35c586160a09b2ae9364f3543c4534"} Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.870844 4771 scope.go:117] "RemoveContainer" containerID="affa16a74eb0604b86c3314a14ed76c9737a8b909faa51015aa11f11a5e5ed87" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.871350 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xmmr" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.896063 4771 scope.go:117] "RemoveContainer" containerID="8f2f83868c1a1f8052fc7665e25e6c04eb62225f5b33b6b064df7df2fda81daf" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.912236 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7xmmr"] Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.921810 4771 scope.go:117] "RemoveContainer" containerID="c5a329e0e5cf44e57bffbf6d9a5c1b911ffedcbcaf5d241195fe4b896e864099" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.925805 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7xmmr"] Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.977195 4771 scope.go:117] "RemoveContainer" containerID="affa16a74eb0604b86c3314a14ed76c9737a8b909faa51015aa11f11a5e5ed87" Oct 02 11:37:04 crc kubenswrapper[4771]: E1002 11:37:04.978354 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"affa16a74eb0604b86c3314a14ed76c9737a8b909faa51015aa11f11a5e5ed87\": container with ID starting with affa16a74eb0604b86c3314a14ed76c9737a8b909faa51015aa11f11a5e5ed87 not found: ID does not exist" containerID="affa16a74eb0604b86c3314a14ed76c9737a8b909faa51015aa11f11a5e5ed87" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.978392 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"affa16a74eb0604b86c3314a14ed76c9737a8b909faa51015aa11f11a5e5ed87"} err="failed to get container status \"affa16a74eb0604b86c3314a14ed76c9737a8b909faa51015aa11f11a5e5ed87\": rpc error: code = NotFound desc = could not find container \"affa16a74eb0604b86c3314a14ed76c9737a8b909faa51015aa11f11a5e5ed87\": container with ID starting with affa16a74eb0604b86c3314a14ed76c9737a8b909faa51015aa11f11a5e5ed87 not found: ID does not exist" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.978419 4771 scope.go:117] "RemoveContainer" containerID="8f2f83868c1a1f8052fc7665e25e6c04eb62225f5b33b6b064df7df2fda81daf" Oct 02 11:37:04 crc kubenswrapper[4771]: E1002 11:37:04.980532 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f2f83868c1a1f8052fc7665e25e6c04eb62225f5b33b6b064df7df2fda81daf\": container with ID starting with 8f2f83868c1a1f8052fc7665e25e6c04eb62225f5b33b6b064df7df2fda81daf not found: ID does not exist" containerID="8f2f83868c1a1f8052fc7665e25e6c04eb62225f5b33b6b064df7df2fda81daf" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.980679 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f2f83868c1a1f8052fc7665e25e6c04eb62225f5b33b6b064df7df2fda81daf"} err="failed to get container status \"8f2f83868c1a1f8052fc7665e25e6c04eb62225f5b33b6b064df7df2fda81daf\": rpc error: code = NotFound desc = could not find container \"8f2f83868c1a1f8052fc7665e25e6c04eb62225f5b33b6b064df7df2fda81daf\": container with ID starting with 8f2f83868c1a1f8052fc7665e25e6c04eb62225f5b33b6b064df7df2fda81daf not found: ID does not exist" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.980767 4771 scope.go:117] "RemoveContainer" containerID="c5a329e0e5cf44e57bffbf6d9a5c1b911ffedcbcaf5d241195fe4b896e864099" Oct 02 11:37:04 crc kubenswrapper[4771]: E1002 11:37:04.981150 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5a329e0e5cf44e57bffbf6d9a5c1b911ffedcbcaf5d241195fe4b896e864099\": container with ID starting with c5a329e0e5cf44e57bffbf6d9a5c1b911ffedcbcaf5d241195fe4b896e864099 not found: ID does not exist" containerID="c5a329e0e5cf44e57bffbf6d9a5c1b911ffedcbcaf5d241195fe4b896e864099" Oct 02 11:37:04 crc kubenswrapper[4771]: I1002 11:37:04.981255 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a329e0e5cf44e57bffbf6d9a5c1b911ffedcbcaf5d241195fe4b896e864099"} err="failed to get container status \"c5a329e0e5cf44e57bffbf6d9a5c1b911ffedcbcaf5d241195fe4b896e864099\": rpc error: code = NotFound desc = could not find container \"c5a329e0e5cf44e57bffbf6d9a5c1b911ffedcbcaf5d241195fe4b896e864099\": container with ID starting with c5a329e0e5cf44e57bffbf6d9a5c1b911ffedcbcaf5d241195fe4b896e864099 not found: ID does not exist" Oct 02 11:37:05 crc kubenswrapper[4771]: I1002 11:37:05.697770 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a70441f1-e293-4c49-8206-1c1806927941" path="/var/lib/kubelet/pods/a70441f1-e293-4c49-8206-1c1806927941/volumes" Oct 02 11:37:09 crc kubenswrapper[4771]: I1002 11:37:09.682637 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:37:09 crc kubenswrapper[4771]: E1002 11:37:09.683642 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:37:24 crc kubenswrapper[4771]: I1002 11:37:24.682320 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:37:24 crc kubenswrapper[4771]: E1002 11:37:24.683656 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:37:35 crc kubenswrapper[4771]: I1002 11:37:35.681785 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:37:35 crc kubenswrapper[4771]: E1002 11:37:35.682708 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:37:48 crc kubenswrapper[4771]: I1002 11:37:48.681399 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:37:48 crc kubenswrapper[4771]: E1002 11:37:48.683084 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:38:01 crc kubenswrapper[4771]: I1002 11:38:01.682523 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:38:01 crc kubenswrapper[4771]: E1002 11:38:01.683662 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:38:07 crc kubenswrapper[4771]: I1002 11:38:07.459980 4771 scope.go:117] "RemoveContainer" containerID="d7a866ae2f65e21389ea5e9018e9e6931c90d3e893268bee302e5cd9e0b0d70b" Oct 02 11:38:07 crc kubenswrapper[4771]: I1002 11:38:07.508924 4771 scope.go:117] "RemoveContainer" containerID="a95ae9a1a1a4526c53f0a214967cace838c073d1909448270027bea58f92a8aa" Oct 02 11:38:07 crc kubenswrapper[4771]: I1002 11:38:07.562066 4771 scope.go:117] "RemoveContainer" containerID="d2e4be0ba2bbbc511e8143d26b7540cf7ded6b3aa3b697ab17595359bea52a2d" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.103051 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rlf2j"] Oct 02 11:38:09 crc kubenswrapper[4771]: E1002 11:38:09.103848 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70441f1-e293-4c49-8206-1c1806927941" containerName="extract-content" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.103865 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70441f1-e293-4c49-8206-1c1806927941" containerName="extract-content" Oct 02 11:38:09 crc kubenswrapper[4771]: E1002 11:38:09.103892 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70441f1-e293-4c49-8206-1c1806927941" containerName="extract-utilities" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.103899 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70441f1-e293-4c49-8206-1c1806927941" containerName="extract-utilities" Oct 02 11:38:09 crc kubenswrapper[4771]: E1002 11:38:09.103915 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70441f1-e293-4c49-8206-1c1806927941" containerName="registry-server" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.103921 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70441f1-e293-4c49-8206-1c1806927941" containerName="registry-server" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.104209 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a70441f1-e293-4c49-8206-1c1806927941" containerName="registry-server" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.105937 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.115947 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rlf2j"] Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.208273 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-catalog-content\") pod \"community-operators-rlf2j\" (UID: \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\") " pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.208344 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-utilities\") pod \"community-operators-rlf2j\" (UID: \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\") " pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.208428 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db9nz\" (UniqueName: \"kubernetes.io/projected/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-kube-api-access-db9nz\") pod \"community-operators-rlf2j\" (UID: \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\") " pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.312188 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db9nz\" (UniqueName: \"kubernetes.io/projected/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-kube-api-access-db9nz\") pod \"community-operators-rlf2j\" (UID: \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\") " pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.312684 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-catalog-content\") pod \"community-operators-rlf2j\" (UID: \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\") " pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.312790 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-utilities\") pod \"community-operators-rlf2j\" (UID: \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\") " pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.313709 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-utilities\") pod \"community-operators-rlf2j\" (UID: \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\") " pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.313784 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-catalog-content\") pod \"community-operators-rlf2j\" (UID: \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\") " pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.348832 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db9nz\" (UniqueName: \"kubernetes.io/projected/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-kube-api-access-db9nz\") pod \"community-operators-rlf2j\" (UID: \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\") " pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:09 crc kubenswrapper[4771]: I1002 11:38:09.434768 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:10 crc kubenswrapper[4771]: I1002 11:38:10.169936 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rlf2j"] Oct 02 11:38:10 crc kubenswrapper[4771]: I1002 11:38:10.672560 4771 generic.go:334] "Generic (PLEG): container finished" podID="f665cc15-4c0e-45c9-b1e1-9faaca0000c3" containerID="b1b02dba0c204955c89b53a6cd133b8183e67ac1d3f1a04f8b8ab854bcac7bff" exitCode=0 Oct 02 11:38:10 crc kubenswrapper[4771]: I1002 11:38:10.672862 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlf2j" event={"ID":"f665cc15-4c0e-45c9-b1e1-9faaca0000c3","Type":"ContainerDied","Data":"b1b02dba0c204955c89b53a6cd133b8183e67ac1d3f1a04f8b8ab854bcac7bff"} Oct 02 11:38:10 crc kubenswrapper[4771]: I1002 11:38:10.672900 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlf2j" event={"ID":"f665cc15-4c0e-45c9-b1e1-9faaca0000c3","Type":"ContainerStarted","Data":"91d90236b03317298644a03d25304cbc0ea5762ae73debf8e86a1abf5539c572"} Oct 02 11:38:12 crc kubenswrapper[4771]: I1002 11:38:12.710534 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlf2j" event={"ID":"f665cc15-4c0e-45c9-b1e1-9faaca0000c3","Type":"ContainerStarted","Data":"64ec96cfbdcfd553c58d753eb88e579b36aa4ef2f2019aea2268ffede9ca26a8"} Oct 02 11:38:14 crc kubenswrapper[4771]: I1002 11:38:14.764930 4771 generic.go:334] "Generic (PLEG): container finished" podID="f665cc15-4c0e-45c9-b1e1-9faaca0000c3" containerID="64ec96cfbdcfd553c58d753eb88e579b36aa4ef2f2019aea2268ffede9ca26a8" exitCode=0 Oct 02 11:38:14 crc kubenswrapper[4771]: I1002 11:38:14.765042 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlf2j" event={"ID":"f665cc15-4c0e-45c9-b1e1-9faaca0000c3","Type":"ContainerDied","Data":"64ec96cfbdcfd553c58d753eb88e579b36aa4ef2f2019aea2268ffede9ca26a8"} Oct 02 11:38:14 crc kubenswrapper[4771]: I1002 11:38:14.770373 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:38:15 crc kubenswrapper[4771]: I1002 11:38:15.682174 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:38:15 crc kubenswrapper[4771]: E1002 11:38:15.682976 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:38:15 crc kubenswrapper[4771]: I1002 11:38:15.778601 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlf2j" event={"ID":"f665cc15-4c0e-45c9-b1e1-9faaca0000c3","Type":"ContainerStarted","Data":"20be41fd7d9d03d06bdc7d7626740fc68289635538c3e8752dff11402627ac72"} Oct 02 11:38:15 crc kubenswrapper[4771]: I1002 11:38:15.812210 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rlf2j" podStartSLOduration=2.279850029 podStartE2EDuration="6.812190791s" podCreationTimestamp="2025-10-02 11:38:09 +0000 UTC" firstStartedPulling="2025-10-02 11:38:10.67870209 +0000 UTC m=+7278.326387157" lastFinishedPulling="2025-10-02 11:38:15.211042852 +0000 UTC m=+7282.858727919" observedRunningTime="2025-10-02 11:38:15.806778911 +0000 UTC m=+7283.454463988" watchObservedRunningTime="2025-10-02 11:38:15.812190791 +0000 UTC m=+7283.459875858" Oct 02 11:38:19 crc kubenswrapper[4771]: I1002 11:38:19.435626 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:19 crc kubenswrapper[4771]: I1002 11:38:19.438719 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:19 crc kubenswrapper[4771]: I1002 11:38:19.494798 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:20 crc kubenswrapper[4771]: I1002 11:38:20.899999 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:20 crc kubenswrapper[4771]: I1002 11:38:20.966505 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rlf2j"] Oct 02 11:38:22 crc kubenswrapper[4771]: I1002 11:38:22.858766 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rlf2j" podUID="f665cc15-4c0e-45c9-b1e1-9faaca0000c3" containerName="registry-server" containerID="cri-o://20be41fd7d9d03d06bdc7d7626740fc68289635538c3e8752dff11402627ac72" gracePeriod=2 Oct 02 11:38:23 crc kubenswrapper[4771]: I1002 11:38:23.880014 4771 generic.go:334] "Generic (PLEG): container finished" podID="f665cc15-4c0e-45c9-b1e1-9faaca0000c3" containerID="20be41fd7d9d03d06bdc7d7626740fc68289635538c3e8752dff11402627ac72" exitCode=0 Oct 02 11:38:23 crc kubenswrapper[4771]: I1002 11:38:23.880632 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlf2j" event={"ID":"f665cc15-4c0e-45c9-b1e1-9faaca0000c3","Type":"ContainerDied","Data":"20be41fd7d9d03d06bdc7d7626740fc68289635538c3e8752dff11402627ac72"} Oct 02 11:38:23 crc kubenswrapper[4771]: I1002 11:38:23.880722 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlf2j" event={"ID":"f665cc15-4c0e-45c9-b1e1-9faaca0000c3","Type":"ContainerDied","Data":"91d90236b03317298644a03d25304cbc0ea5762ae73debf8e86a1abf5539c572"} Oct 02 11:38:23 crc kubenswrapper[4771]: I1002 11:38:23.880736 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91d90236b03317298644a03d25304cbc0ea5762ae73debf8e86a1abf5539c572" Oct 02 11:38:23 crc kubenswrapper[4771]: I1002 11:38:23.948199 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:24 crc kubenswrapper[4771]: I1002 11:38:24.043475 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db9nz\" (UniqueName: \"kubernetes.io/projected/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-kube-api-access-db9nz\") pod \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\" (UID: \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\") " Oct 02 11:38:24 crc kubenswrapper[4771]: I1002 11:38:24.043611 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-utilities\") pod \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\" (UID: \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\") " Oct 02 11:38:24 crc kubenswrapper[4771]: I1002 11:38:24.043757 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-catalog-content\") pod \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\" (UID: \"f665cc15-4c0e-45c9-b1e1-9faaca0000c3\") " Oct 02 11:38:24 crc kubenswrapper[4771]: I1002 11:38:24.044664 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-utilities" (OuterVolumeSpecName: "utilities") pod "f665cc15-4c0e-45c9-b1e1-9faaca0000c3" (UID: "f665cc15-4c0e-45c9-b1e1-9faaca0000c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:38:24 crc kubenswrapper[4771]: I1002 11:38:24.067685 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-kube-api-access-db9nz" (OuterVolumeSpecName: "kube-api-access-db9nz") pod "f665cc15-4c0e-45c9-b1e1-9faaca0000c3" (UID: "f665cc15-4c0e-45c9-b1e1-9faaca0000c3"). InnerVolumeSpecName "kube-api-access-db9nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:38:24 crc kubenswrapper[4771]: I1002 11:38:24.094145 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f665cc15-4c0e-45c9-b1e1-9faaca0000c3" (UID: "f665cc15-4c0e-45c9-b1e1-9faaca0000c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:38:24 crc kubenswrapper[4771]: I1002 11:38:24.146968 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db9nz\" (UniqueName: \"kubernetes.io/projected/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-kube-api-access-db9nz\") on node \"crc\" DevicePath \"\"" Oct 02 11:38:24 crc kubenswrapper[4771]: I1002 11:38:24.147005 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:38:24 crc kubenswrapper[4771]: I1002 11:38:24.147014 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f665cc15-4c0e-45c9-b1e1-9faaca0000c3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:38:24 crc kubenswrapper[4771]: I1002 11:38:24.888717 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlf2j" Oct 02 11:38:24 crc kubenswrapper[4771]: I1002 11:38:24.929254 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rlf2j"] Oct 02 11:38:24 crc kubenswrapper[4771]: I1002 11:38:24.941068 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rlf2j"] Oct 02 11:38:25 crc kubenswrapper[4771]: I1002 11:38:25.704063 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f665cc15-4c0e-45c9-b1e1-9faaca0000c3" path="/var/lib/kubelet/pods/f665cc15-4c0e-45c9-b1e1-9faaca0000c3/volumes" Oct 02 11:38:26 crc kubenswrapper[4771]: I1002 11:38:26.682394 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:38:26 crc kubenswrapper[4771]: E1002 11:38:26.683145 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:38:40 crc kubenswrapper[4771]: I1002 11:38:40.682330 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:38:40 crc kubenswrapper[4771]: E1002 11:38:40.684671 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.400920 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 02 11:38:42 crc kubenswrapper[4771]: E1002 11:38:42.402549 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f665cc15-4c0e-45c9-b1e1-9faaca0000c3" containerName="extract-utilities" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.402564 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f665cc15-4c0e-45c9-b1e1-9faaca0000c3" containerName="extract-utilities" Oct 02 11:38:42 crc kubenswrapper[4771]: E1002 11:38:42.402585 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f665cc15-4c0e-45c9-b1e1-9faaca0000c3" containerName="registry-server" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.402591 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f665cc15-4c0e-45c9-b1e1-9faaca0000c3" containerName="registry-server" Oct 02 11:38:42 crc kubenswrapper[4771]: E1002 11:38:42.402616 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f665cc15-4c0e-45c9-b1e1-9faaca0000c3" containerName="extract-content" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.402623 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="f665cc15-4c0e-45c9-b1e1-9faaca0000c3" containerName="extract-content" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.402836 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="f665cc15-4c0e-45c9-b1e1-9faaca0000c3" containerName="registry-server" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.403742 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.407742 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.407978 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.408015 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.408313 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t4pv9" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.421845 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.545270 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57b7360c-088e-4c95-a10b-e7d034d2f83f-config-data\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.545353 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/57b7360c-088e-4c95-a10b-e7d034d2f83f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.545421 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.545481 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc2vc\" (UniqueName: \"kubernetes.io/projected/57b7360c-088e-4c95-a10b-e7d034d2f83f-kube-api-access-vc2vc\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.545592 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.545645 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.545675 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/57b7360c-088e-4c95-a10b-e7d034d2f83f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.545735 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.545761 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/57b7360c-088e-4c95-a10b-e7d034d2f83f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.647603 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc2vc\" (UniqueName: \"kubernetes.io/projected/57b7360c-088e-4c95-a10b-e7d034d2f83f-kube-api-access-vc2vc\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.647723 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.647766 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.647793 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/57b7360c-088e-4c95-a10b-e7d034d2f83f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.647832 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.647854 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/57b7360c-088e-4c95-a10b-e7d034d2f83f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.647919 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57b7360c-088e-4c95-a10b-e7d034d2f83f-config-data\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.647965 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/57b7360c-088e-4c95-a10b-e7d034d2f83f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.648005 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.648645 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/57b7360c-088e-4c95-a10b-e7d034d2f83f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.649149 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/57b7360c-088e-4c95-a10b-e7d034d2f83f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.649349 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57b7360c-088e-4c95-a10b-e7d034d2f83f-config-data\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.649626 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/57b7360c-088e-4c95-a10b-e7d034d2f83f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.650237 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.655651 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.655809 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.666790 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.670267 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc2vc\" (UniqueName: \"kubernetes.io/projected/57b7360c-088e-4c95-a10b-e7d034d2f83f-kube-api-access-vc2vc\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.686495 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " pod="openstack/tempest-tests-tempest" Oct 02 11:38:42 crc kubenswrapper[4771]: I1002 11:38:42.742770 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 11:38:43 crc kubenswrapper[4771]: I1002 11:38:43.240377 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 02 11:38:44 crc kubenswrapper[4771]: I1002 11:38:44.130120 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"57b7360c-088e-4c95-a10b-e7d034d2f83f","Type":"ContainerStarted","Data":"b293abd59af9ef0213e59fe9152a108d8eeafd674b316be8cf0632416ead92ad"} Oct 02 11:38:52 crc kubenswrapper[4771]: I1002 11:38:52.681564 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:38:52 crc kubenswrapper[4771]: E1002 11:38:52.682474 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:39:06 crc kubenswrapper[4771]: I1002 11:39:06.682932 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:39:06 crc kubenswrapper[4771]: E1002 11:39:06.683786 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:39:18 crc kubenswrapper[4771]: I1002 11:39:18.681937 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:39:18 crc kubenswrapper[4771]: E1002 11:39:18.685101 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:39:22 crc kubenswrapper[4771]: E1002 11:39:22.265705 4771 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 02 11:39:22 crc kubenswrapper[4771]: E1002 11:39:22.268954 4771 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vc2vc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(57b7360c-088e-4c95-a10b-e7d034d2f83f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:39:22 crc kubenswrapper[4771]: E1002 11:39:22.270325 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="57b7360c-088e-4c95-a10b-e7d034d2f83f" Oct 02 11:39:22 crc kubenswrapper[4771]: E1002 11:39:22.641984 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="57b7360c-088e-4c95-a10b-e7d034d2f83f" Oct 02 11:39:33 crc kubenswrapper[4771]: I1002 11:39:33.699661 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:39:33 crc kubenswrapper[4771]: E1002 11:39:33.701077 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:39:34 crc kubenswrapper[4771]: I1002 11:39:34.106107 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 02 11:39:35 crc kubenswrapper[4771]: I1002 11:39:35.897227 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"57b7360c-088e-4c95-a10b-e7d034d2f83f","Type":"ContainerStarted","Data":"a291cba449d68e66966b82fad2d4d0240e3fda98f1e27cfbca19bbffd84391e9"} Oct 02 11:39:35 crc kubenswrapper[4771]: I1002 11:39:35.919002 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.065606889 podStartE2EDuration="54.918984323s" podCreationTimestamp="2025-10-02 11:38:41 +0000 UTC" firstStartedPulling="2025-10-02 11:38:43.250384479 +0000 UTC m=+7310.898069546" lastFinishedPulling="2025-10-02 11:39:34.103761923 +0000 UTC m=+7361.751446980" observedRunningTime="2025-10-02 11:39:35.912802944 +0000 UTC m=+7363.560488021" watchObservedRunningTime="2025-10-02 11:39:35.918984323 +0000 UTC m=+7363.566669390" Oct 02 11:39:46 crc kubenswrapper[4771]: I1002 11:39:46.682529 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:39:46 crc kubenswrapper[4771]: E1002 11:39:46.683474 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:39:58 crc kubenswrapper[4771]: I1002 11:39:58.682293 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:39:58 crc kubenswrapper[4771]: E1002 11:39:58.683592 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:40:10 crc kubenswrapper[4771]: I1002 11:40:10.681319 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:40:10 crc kubenswrapper[4771]: E1002 11:40:10.682351 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:40:22 crc kubenswrapper[4771]: I1002 11:40:22.681484 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:40:23 crc kubenswrapper[4771]: I1002 11:40:23.439896 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"3bb74c9f46be75efdbfc800b6d2f38f7da79e737ae4bf6020f5528c548db4dc6"} Oct 02 11:40:41 crc kubenswrapper[4771]: I1002 11:40:41.654096 4771 generic.go:334] "Generic (PLEG): container finished" podID="57b7360c-088e-4c95-a10b-e7d034d2f83f" containerID="a291cba449d68e66966b82fad2d4d0240e3fda98f1e27cfbca19bbffd84391e9" exitCode=123 Oct 02 11:40:41 crc kubenswrapper[4771]: I1002 11:40:41.654235 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"57b7360c-088e-4c95-a10b-e7d034d2f83f","Type":"ContainerDied","Data":"a291cba449d68e66966b82fad2d4d0240e3fda98f1e27cfbca19bbffd84391e9"} Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.093695 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.150414 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vc2vc\" (UniqueName: \"kubernetes.io/projected/57b7360c-088e-4c95-a10b-e7d034d2f83f-kube-api-access-vc2vc\") pod \"57b7360c-088e-4c95-a10b-e7d034d2f83f\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.151248 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/57b7360c-088e-4c95-a10b-e7d034d2f83f-test-operator-ephemeral-workdir\") pod \"57b7360c-088e-4c95-a10b-e7d034d2f83f\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.151331 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/57b7360c-088e-4c95-a10b-e7d034d2f83f-test-operator-ephemeral-temporary\") pod \"57b7360c-088e-4c95-a10b-e7d034d2f83f\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.151469 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-ssh-key\") pod \"57b7360c-088e-4c95-a10b-e7d034d2f83f\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.151554 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"57b7360c-088e-4c95-a10b-e7d034d2f83f\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.151608 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57b7360c-088e-4c95-a10b-e7d034d2f83f-config-data\") pod \"57b7360c-088e-4c95-a10b-e7d034d2f83f\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.151702 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/57b7360c-088e-4c95-a10b-e7d034d2f83f-openstack-config\") pod \"57b7360c-088e-4c95-a10b-e7d034d2f83f\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.151751 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-openstack-config-secret\") pod \"57b7360c-088e-4c95-a10b-e7d034d2f83f\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.151942 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-ca-certs\") pod \"57b7360c-088e-4c95-a10b-e7d034d2f83f\" (UID: \"57b7360c-088e-4c95-a10b-e7d034d2f83f\") " Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.152096 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57b7360c-088e-4c95-a10b-e7d034d2f83f-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "57b7360c-088e-4c95-a10b-e7d034d2f83f" (UID: "57b7360c-088e-4c95-a10b-e7d034d2f83f"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.153295 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57b7360c-088e-4c95-a10b-e7d034d2f83f-config-data" (OuterVolumeSpecName: "config-data") pod "57b7360c-088e-4c95-a10b-e7d034d2f83f" (UID: "57b7360c-088e-4c95-a10b-e7d034d2f83f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.153832 4771 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/57b7360c-088e-4c95-a10b-e7d034d2f83f-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.153860 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57b7360c-088e-4c95-a10b-e7d034d2f83f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.155442 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57b7360c-088e-4c95-a10b-e7d034d2f83f-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "57b7360c-088e-4c95-a10b-e7d034d2f83f" (UID: "57b7360c-088e-4c95-a10b-e7d034d2f83f"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.158999 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "57b7360c-088e-4c95-a10b-e7d034d2f83f" (UID: "57b7360c-088e-4c95-a10b-e7d034d2f83f"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.159220 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57b7360c-088e-4c95-a10b-e7d034d2f83f-kube-api-access-vc2vc" (OuterVolumeSpecName: "kube-api-access-vc2vc") pod "57b7360c-088e-4c95-a10b-e7d034d2f83f" (UID: "57b7360c-088e-4c95-a10b-e7d034d2f83f"). InnerVolumeSpecName "kube-api-access-vc2vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.188043 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "57b7360c-088e-4c95-a10b-e7d034d2f83f" (UID: "57b7360c-088e-4c95-a10b-e7d034d2f83f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.188088 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "57b7360c-088e-4c95-a10b-e7d034d2f83f" (UID: "57b7360c-088e-4c95-a10b-e7d034d2f83f"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.200472 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "57b7360c-088e-4c95-a10b-e7d034d2f83f" (UID: "57b7360c-088e-4c95-a10b-e7d034d2f83f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.220465 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57b7360c-088e-4c95-a10b-e7d034d2f83f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "57b7360c-088e-4c95-a10b-e7d034d2f83f" (UID: "57b7360c-088e-4c95-a10b-e7d034d2f83f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.258293 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vc2vc\" (UniqueName: \"kubernetes.io/projected/57b7360c-088e-4c95-a10b-e7d034d2f83f-kube-api-access-vc2vc\") on node \"crc\" DevicePath \"\"" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.258356 4771 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/57b7360c-088e-4c95-a10b-e7d034d2f83f-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.258378 4771 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.258992 4771 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.259022 4771 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/57b7360c-088e-4c95-a10b-e7d034d2f83f-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.263606 4771 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.263640 4771 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/57b7360c-088e-4c95-a10b-e7d034d2f83f-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.296419 4771 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.366325 4771 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.679675 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"57b7360c-088e-4c95-a10b-e7d034d2f83f","Type":"ContainerDied","Data":"b293abd59af9ef0213e59fe9152a108d8eeafd674b316be8cf0632416ead92ad"} Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.679854 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b293abd59af9ef0213e59fe9152a108d8eeafd674b316be8cf0632416ead92ad" Oct 02 11:40:43 crc kubenswrapper[4771]: I1002 11:40:43.680005 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.656827 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 11:40:53 crc kubenswrapper[4771]: E1002 11:40:53.657915 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b7360c-088e-4c95-a10b-e7d034d2f83f" containerName="tempest-tests-tempest-tests-runner" Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.657932 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b7360c-088e-4c95-a10b-e7d034d2f83f" containerName="tempest-tests-tempest-tests-runner" Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.658265 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b7360c-088e-4c95-a10b-e7d034d2f83f" containerName="tempest-tests-tempest-tests-runner" Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.659624 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.662082 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t4pv9" Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.701595 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.725693 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpbxx\" (UniqueName: \"kubernetes.io/projected/3e55f1c1-c148-4ee4-80fb-29c6029f2d2d-kube-api-access-jpbxx\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3e55f1c1-c148-4ee4-80fb-29c6029f2d2d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.726027 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3e55f1c1-c148-4ee4-80fb-29c6029f2d2d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.828705 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpbxx\" (UniqueName: \"kubernetes.io/projected/3e55f1c1-c148-4ee4-80fb-29c6029f2d2d-kube-api-access-jpbxx\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3e55f1c1-c148-4ee4-80fb-29c6029f2d2d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.828795 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3e55f1c1-c148-4ee4-80fb-29c6029f2d2d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.829311 4771 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3e55f1c1-c148-4ee4-80fb-29c6029f2d2d\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.857790 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpbxx\" (UniqueName: \"kubernetes.io/projected/3e55f1c1-c148-4ee4-80fb-29c6029f2d2d-kube-api-access-jpbxx\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3e55f1c1-c148-4ee4-80fb-29c6029f2d2d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.860177 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3e55f1c1-c148-4ee4-80fb-29c6029f2d2d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 11:40:53 crc kubenswrapper[4771]: I1002 11:40:53.992462 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 11:40:54 crc kubenswrapper[4771]: I1002 11:40:54.507800 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 11:40:54 crc kubenswrapper[4771]: I1002 11:40:54.795712 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"3e55f1c1-c148-4ee4-80fb-29c6029f2d2d","Type":"ContainerStarted","Data":"87f387b5655afe4f762262eaf9ed7538aa28ec7adebe6756c0f7b15ad5c23cd7"} Oct 02 11:40:55 crc kubenswrapper[4771]: I1002 11:40:55.810886 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"3e55f1c1-c148-4ee4-80fb-29c6029f2d2d","Type":"ContainerStarted","Data":"c34e7c113c9e49df6c9dc9c99dbb3d6d13333ba8313008a41c70474e8cf6befa"} Oct 02 11:40:55 crc kubenswrapper[4771]: I1002 11:40:55.837060 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.011026674 podStartE2EDuration="2.837041566s" podCreationTimestamp="2025-10-02 11:40:53 +0000 UTC" firstStartedPulling="2025-10-02 11:40:54.516893331 +0000 UTC m=+7442.164578398" lastFinishedPulling="2025-10-02 11:40:55.342908223 +0000 UTC m=+7442.990593290" observedRunningTime="2025-10-02 11:40:55.828868925 +0000 UTC m=+7443.476554002" watchObservedRunningTime="2025-10-02 11:40:55.837041566 +0000 UTC m=+7443.484726633" Oct 02 11:41:22 crc kubenswrapper[4771]: I1002 11:41:22.192983 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pc9ng/must-gather-j92sz"] Oct 02 11:41:22 crc kubenswrapper[4771]: I1002 11:41:22.196332 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/must-gather-j92sz" Oct 02 11:41:22 crc kubenswrapper[4771]: I1002 11:41:22.199672 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pc9ng"/"kube-root-ca.crt" Oct 02 11:41:22 crc kubenswrapper[4771]: I1002 11:41:22.199723 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pc9ng"/"openshift-service-ca.crt" Oct 02 11:41:22 crc kubenswrapper[4771]: I1002 11:41:22.200187 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-pc9ng"/"default-dockercfg-dhrmc" Oct 02 11:41:22 crc kubenswrapper[4771]: I1002 11:41:22.210157 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pc9ng/must-gather-j92sz"] Oct 02 11:41:22 crc kubenswrapper[4771]: I1002 11:41:22.259059 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e26af54-9be2-44da-9bbd-0684c864ce9e-must-gather-output\") pod \"must-gather-j92sz\" (UID: \"1e26af54-9be2-44da-9bbd-0684c864ce9e\") " pod="openshift-must-gather-pc9ng/must-gather-j92sz" Oct 02 11:41:22 crc kubenswrapper[4771]: I1002 11:41:22.259271 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws8fg\" (UniqueName: \"kubernetes.io/projected/1e26af54-9be2-44da-9bbd-0684c864ce9e-kube-api-access-ws8fg\") pod \"must-gather-j92sz\" (UID: \"1e26af54-9be2-44da-9bbd-0684c864ce9e\") " pod="openshift-must-gather-pc9ng/must-gather-j92sz" Oct 02 11:41:22 crc kubenswrapper[4771]: I1002 11:41:22.361832 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e26af54-9be2-44da-9bbd-0684c864ce9e-must-gather-output\") pod \"must-gather-j92sz\" (UID: \"1e26af54-9be2-44da-9bbd-0684c864ce9e\") " pod="openshift-must-gather-pc9ng/must-gather-j92sz" Oct 02 11:41:22 crc kubenswrapper[4771]: I1002 11:41:22.361939 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws8fg\" (UniqueName: \"kubernetes.io/projected/1e26af54-9be2-44da-9bbd-0684c864ce9e-kube-api-access-ws8fg\") pod \"must-gather-j92sz\" (UID: \"1e26af54-9be2-44da-9bbd-0684c864ce9e\") " pod="openshift-must-gather-pc9ng/must-gather-j92sz" Oct 02 11:41:22 crc kubenswrapper[4771]: I1002 11:41:22.362419 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e26af54-9be2-44da-9bbd-0684c864ce9e-must-gather-output\") pod \"must-gather-j92sz\" (UID: \"1e26af54-9be2-44da-9bbd-0684c864ce9e\") " pod="openshift-must-gather-pc9ng/must-gather-j92sz" Oct 02 11:41:22 crc kubenswrapper[4771]: I1002 11:41:22.385260 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws8fg\" (UniqueName: \"kubernetes.io/projected/1e26af54-9be2-44da-9bbd-0684c864ce9e-kube-api-access-ws8fg\") pod \"must-gather-j92sz\" (UID: \"1e26af54-9be2-44da-9bbd-0684c864ce9e\") " pod="openshift-must-gather-pc9ng/must-gather-j92sz" Oct 02 11:41:22 crc kubenswrapper[4771]: I1002 11:41:22.521380 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/must-gather-j92sz" Oct 02 11:41:23 crc kubenswrapper[4771]: I1002 11:41:23.055663 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pc9ng/must-gather-j92sz"] Oct 02 11:41:23 crc kubenswrapper[4771]: I1002 11:41:23.142936 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pc9ng/must-gather-j92sz" event={"ID":"1e26af54-9be2-44da-9bbd-0684c864ce9e","Type":"ContainerStarted","Data":"ad92f191061cdd447aa48134eede19edf0b2a667fd5ecacc6bd71e71f34c0a51"} Oct 02 11:41:33 crc kubenswrapper[4771]: I1002 11:41:33.279804 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pc9ng/must-gather-j92sz" event={"ID":"1e26af54-9be2-44da-9bbd-0684c864ce9e","Type":"ContainerStarted","Data":"30e6fe3afab65597caa31dc9566ca49b39d65338a78357367c6cb3c4003713ee"} Oct 02 11:41:33 crc kubenswrapper[4771]: I1002 11:41:33.280387 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pc9ng/must-gather-j92sz" event={"ID":"1e26af54-9be2-44da-9bbd-0684c864ce9e","Type":"ContainerStarted","Data":"21e072d36ab3899659bc92f035fc77bf6c5e73d35b8d61fa026f411d0baf4ee0"} Oct 02 11:41:33 crc kubenswrapper[4771]: I1002 11:41:33.300372 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pc9ng/must-gather-j92sz" podStartSLOduration=2.127348787 podStartE2EDuration="11.300355327s" podCreationTimestamp="2025-10-02 11:41:22 +0000 UTC" firstStartedPulling="2025-10-02 11:41:23.050938091 +0000 UTC m=+7470.698623158" lastFinishedPulling="2025-10-02 11:41:32.223944631 +0000 UTC m=+7479.871629698" observedRunningTime="2025-10-02 11:41:33.291876568 +0000 UTC m=+7480.939561645" watchObservedRunningTime="2025-10-02 11:41:33.300355327 +0000 UTC m=+7480.948040394" Oct 02 11:41:37 crc kubenswrapper[4771]: I1002 11:41:37.482591 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pc9ng/crc-debug-ks887"] Oct 02 11:41:37 crc kubenswrapper[4771]: I1002 11:41:37.486536 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/crc-debug-ks887" Oct 02 11:41:37 crc kubenswrapper[4771]: I1002 11:41:37.568190 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p74c\" (UniqueName: \"kubernetes.io/projected/85fb7ae5-5b81-4fc7-9527-ab0df063d173-kube-api-access-9p74c\") pod \"crc-debug-ks887\" (UID: \"85fb7ae5-5b81-4fc7-9527-ab0df063d173\") " pod="openshift-must-gather-pc9ng/crc-debug-ks887" Oct 02 11:41:37 crc kubenswrapper[4771]: I1002 11:41:37.568268 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/85fb7ae5-5b81-4fc7-9527-ab0df063d173-host\") pod \"crc-debug-ks887\" (UID: \"85fb7ae5-5b81-4fc7-9527-ab0df063d173\") " pod="openshift-must-gather-pc9ng/crc-debug-ks887" Oct 02 11:41:37 crc kubenswrapper[4771]: I1002 11:41:37.670446 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p74c\" (UniqueName: \"kubernetes.io/projected/85fb7ae5-5b81-4fc7-9527-ab0df063d173-kube-api-access-9p74c\") pod \"crc-debug-ks887\" (UID: \"85fb7ae5-5b81-4fc7-9527-ab0df063d173\") " pod="openshift-must-gather-pc9ng/crc-debug-ks887" Oct 02 11:41:37 crc kubenswrapper[4771]: I1002 11:41:37.670530 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/85fb7ae5-5b81-4fc7-9527-ab0df063d173-host\") pod \"crc-debug-ks887\" (UID: \"85fb7ae5-5b81-4fc7-9527-ab0df063d173\") " pod="openshift-must-gather-pc9ng/crc-debug-ks887" Oct 02 11:41:37 crc kubenswrapper[4771]: I1002 11:41:37.670866 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/85fb7ae5-5b81-4fc7-9527-ab0df063d173-host\") pod \"crc-debug-ks887\" (UID: \"85fb7ae5-5b81-4fc7-9527-ab0df063d173\") " pod="openshift-must-gather-pc9ng/crc-debug-ks887" Oct 02 11:41:37 crc kubenswrapper[4771]: I1002 11:41:37.701147 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p74c\" (UniqueName: \"kubernetes.io/projected/85fb7ae5-5b81-4fc7-9527-ab0df063d173-kube-api-access-9p74c\") pod \"crc-debug-ks887\" (UID: \"85fb7ae5-5b81-4fc7-9527-ab0df063d173\") " pod="openshift-must-gather-pc9ng/crc-debug-ks887" Oct 02 11:41:37 crc kubenswrapper[4771]: I1002 11:41:37.810236 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/crc-debug-ks887" Oct 02 11:41:38 crc kubenswrapper[4771]: I1002 11:41:38.333977 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pc9ng/crc-debug-ks887" event={"ID":"85fb7ae5-5b81-4fc7-9527-ab0df063d173","Type":"ContainerStarted","Data":"4e2559d672faab69a5113ba43629e005a5f5e4fe8099cc544cb90b68b395a401"} Oct 02 11:41:53 crc kubenswrapper[4771]: I1002 11:41:53.528150 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pc9ng/crc-debug-ks887" event={"ID":"85fb7ae5-5b81-4fc7-9527-ab0df063d173","Type":"ContainerStarted","Data":"400d3412c9eea1469ff0b878295f2fcb0f2b31f92d51dfed0f5b3936c8996a9e"} Oct 02 11:41:53 crc kubenswrapper[4771]: I1002 11:41:53.551936 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pc9ng/crc-debug-ks887" podStartSLOduration=1.7180343219999998 podStartE2EDuration="16.551914796s" podCreationTimestamp="2025-10-02 11:41:37 +0000 UTC" firstStartedPulling="2025-10-02 11:41:37.866729837 +0000 UTC m=+7485.514414904" lastFinishedPulling="2025-10-02 11:41:52.700610321 +0000 UTC m=+7500.348295378" observedRunningTime="2025-10-02 11:41:53.544089844 +0000 UTC m=+7501.191774901" watchObservedRunningTime="2025-10-02 11:41:53.551914796 +0000 UTC m=+7501.199599863" Oct 02 11:42:07 crc kubenswrapper[4771]: I1002 11:42:07.242638 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r2bnm"] Oct 02 11:42:07 crc kubenswrapper[4771]: I1002 11:42:07.252730 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:07 crc kubenswrapper[4771]: I1002 11:42:07.270298 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r2bnm"] Oct 02 11:42:07 crc kubenswrapper[4771]: I1002 11:42:07.331794 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhqp9\" (UniqueName: \"kubernetes.io/projected/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-kube-api-access-nhqp9\") pod \"certified-operators-r2bnm\" (UID: \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\") " pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:07 crc kubenswrapper[4771]: I1002 11:42:07.331966 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-utilities\") pod \"certified-operators-r2bnm\" (UID: \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\") " pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:07 crc kubenswrapper[4771]: I1002 11:42:07.332224 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-catalog-content\") pod \"certified-operators-r2bnm\" (UID: \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\") " pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:07 crc kubenswrapper[4771]: I1002 11:42:07.434960 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-utilities\") pod \"certified-operators-r2bnm\" (UID: \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\") " pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:07 crc kubenswrapper[4771]: I1002 11:42:07.435175 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-catalog-content\") pod \"certified-operators-r2bnm\" (UID: \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\") " pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:07 crc kubenswrapper[4771]: I1002 11:42:07.435215 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhqp9\" (UniqueName: \"kubernetes.io/projected/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-kube-api-access-nhqp9\") pod \"certified-operators-r2bnm\" (UID: \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\") " pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:07 crc kubenswrapper[4771]: I1002 11:42:07.436103 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-utilities\") pod \"certified-operators-r2bnm\" (UID: \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\") " pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:07 crc kubenswrapper[4771]: I1002 11:42:07.436119 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-catalog-content\") pod \"certified-operators-r2bnm\" (UID: \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\") " pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:07 crc kubenswrapper[4771]: I1002 11:42:07.457089 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhqp9\" (UniqueName: \"kubernetes.io/projected/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-kube-api-access-nhqp9\") pod \"certified-operators-r2bnm\" (UID: \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\") " pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:08 crc kubenswrapper[4771]: I1002 11:42:08.993873 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:10 crc kubenswrapper[4771]: W1002 11:42:10.049824 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b9ea4ef_3ad9_4924_a325_036c9d6dd7fe.slice/crio-78bd43209dcbb73841760903dd5f94278c560b415105b57cf208f9d0a63ec7fb WatchSource:0}: Error finding container 78bd43209dcbb73841760903dd5f94278c560b415105b57cf208f9d0a63ec7fb: Status 404 returned error can't find the container with id 78bd43209dcbb73841760903dd5f94278c560b415105b57cf208f9d0a63ec7fb Oct 02 11:42:10 crc kubenswrapper[4771]: I1002 11:42:10.061707 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r2bnm"] Oct 02 11:42:10 crc kubenswrapper[4771]: I1002 11:42:10.794253 4771 generic.go:334] "Generic (PLEG): container finished" podID="9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" containerID="14f89027117ccf5e3a34d8c0d76d067563ed9186804a51a6e49c4f80413677c6" exitCode=0 Oct 02 11:42:10 crc kubenswrapper[4771]: I1002 11:42:10.794442 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2bnm" event={"ID":"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe","Type":"ContainerDied","Data":"14f89027117ccf5e3a34d8c0d76d067563ed9186804a51a6e49c4f80413677c6"} Oct 02 11:42:10 crc kubenswrapper[4771]: I1002 11:42:10.794558 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2bnm" event={"ID":"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe","Type":"ContainerStarted","Data":"78bd43209dcbb73841760903dd5f94278c560b415105b57cf208f9d0a63ec7fb"} Oct 02 11:42:11 crc kubenswrapper[4771]: I1002 11:42:11.808300 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2bnm" event={"ID":"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe","Type":"ContainerStarted","Data":"760bfb63fd2b75f7482faa4b2e098060357063bfd25ddc456eddaff76301326d"} Oct 02 11:42:13 crc kubenswrapper[4771]: I1002 11:42:13.829765 4771 generic.go:334] "Generic (PLEG): container finished" podID="9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" containerID="760bfb63fd2b75f7482faa4b2e098060357063bfd25ddc456eddaff76301326d" exitCode=0 Oct 02 11:42:13 crc kubenswrapper[4771]: I1002 11:42:13.829834 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2bnm" event={"ID":"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe","Type":"ContainerDied","Data":"760bfb63fd2b75f7482faa4b2e098060357063bfd25ddc456eddaff76301326d"} Oct 02 11:42:16 crc kubenswrapper[4771]: I1002 11:42:16.866235 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2bnm" event={"ID":"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe","Type":"ContainerStarted","Data":"6184c31d7f07b226d93618c971b2c43cac96773998783ea6fe74b9b492aff59b"} Oct 02 11:42:16 crc kubenswrapper[4771]: I1002 11:42:16.896800 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r2bnm" podStartSLOduration=4.757343744 podStartE2EDuration="9.896777698s" podCreationTimestamp="2025-10-02 11:42:07 +0000 UTC" firstStartedPulling="2025-10-02 11:42:10.796566478 +0000 UTC m=+7518.444251545" lastFinishedPulling="2025-10-02 11:42:15.936000432 +0000 UTC m=+7523.583685499" observedRunningTime="2025-10-02 11:42:16.891604724 +0000 UTC m=+7524.539289791" watchObservedRunningTime="2025-10-02 11:42:16.896777698 +0000 UTC m=+7524.544462765" Oct 02 11:42:18 crc kubenswrapper[4771]: I1002 11:42:18.995080 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:18 crc kubenswrapper[4771]: I1002 11:42:18.996594 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:19 crc kubenswrapper[4771]: I1002 11:42:19.060187 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:29 crc kubenswrapper[4771]: I1002 11:42:29.178249 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:29 crc kubenswrapper[4771]: I1002 11:42:29.239012 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r2bnm"] Oct 02 11:42:30 crc kubenswrapper[4771]: I1002 11:42:30.055058 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r2bnm" podUID="9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" containerName="registry-server" containerID="cri-o://6184c31d7f07b226d93618c971b2c43cac96773998783ea6fe74b9b492aff59b" gracePeriod=2 Oct 02 11:42:31 crc kubenswrapper[4771]: I1002 11:42:31.069812 4771 generic.go:334] "Generic (PLEG): container finished" podID="9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" containerID="6184c31d7f07b226d93618c971b2c43cac96773998783ea6fe74b9b492aff59b" exitCode=0 Oct 02 11:42:31 crc kubenswrapper[4771]: I1002 11:42:31.070105 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2bnm" event={"ID":"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe","Type":"ContainerDied","Data":"6184c31d7f07b226d93618c971b2c43cac96773998783ea6fe74b9b492aff59b"} Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:32.450643 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:32.614741 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhqp9\" (UniqueName: \"kubernetes.io/projected/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-kube-api-access-nhqp9\") pod \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\" (UID: \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\") " Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:32.614887 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-utilities\") pod \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\" (UID: \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\") " Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:32.615079 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-catalog-content\") pod \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\" (UID: \"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe\") " Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:32.621490 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-utilities" (OuterVolumeSpecName: "utilities") pod "9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" (UID: "9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:32.655835 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-kube-api-access-nhqp9" (OuterVolumeSpecName: "kube-api-access-nhqp9") pod "9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" (UID: "9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe"). InnerVolumeSpecName "kube-api-access-nhqp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:32.717240 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhqp9\" (UniqueName: \"kubernetes.io/projected/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-kube-api-access-nhqp9\") on node \"crc\" DevicePath \"\"" Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:32.717276 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:32.738119 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" (UID: "9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:32.820061 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:33.109416 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2bnm" event={"ID":"9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe","Type":"ContainerDied","Data":"78bd43209dcbb73841760903dd5f94278c560b415105b57cf208f9d0a63ec7fb"} Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:33.109465 4771 scope.go:117] "RemoveContainer" containerID="6184c31d7f07b226d93618c971b2c43cac96773998783ea6fe74b9b492aff59b" Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:33.109511 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2bnm" Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:33.153822 4771 scope.go:117] "RemoveContainer" containerID="760bfb63fd2b75f7482faa4b2e098060357063bfd25ddc456eddaff76301326d" Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:33.171322 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r2bnm"] Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:33.190517 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r2bnm"] Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:33.213826 4771 scope.go:117] "RemoveContainer" containerID="14f89027117ccf5e3a34d8c0d76d067563ed9186804a51a6e49c4f80413677c6" Oct 02 11:42:33 crc kubenswrapper[4771]: I1002 11:42:33.705155 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" path="/var/lib/kubelet/pods/9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe/volumes" Oct 02 11:42:42 crc kubenswrapper[4771]: I1002 11:42:42.146158 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:42:42 crc kubenswrapper[4771]: I1002 11:42:42.147807 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:43:12 crc kubenswrapper[4771]: I1002 11:43:12.146340 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:43:12 crc kubenswrapper[4771]: I1002 11:43:12.146923 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:43:14 crc kubenswrapper[4771]: I1002 11:43:14.625920 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_bc2a08a4-b679-4bcc-ac8a-db9d7108b69c/aodh-api/0.log" Oct 02 11:43:14 crc kubenswrapper[4771]: I1002 11:43:14.755473 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_bc2a08a4-b679-4bcc-ac8a-db9d7108b69c/aodh-evaluator/0.log" Oct 02 11:43:14 crc kubenswrapper[4771]: I1002 11:43:14.873377 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_bc2a08a4-b679-4bcc-ac8a-db9d7108b69c/aodh-listener/0.log" Oct 02 11:43:14 crc kubenswrapper[4771]: I1002 11:43:14.958749 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_bc2a08a4-b679-4bcc-ac8a-db9d7108b69c/aodh-notifier/0.log" Oct 02 11:43:15 crc kubenswrapper[4771]: I1002 11:43:15.152317 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c4746c7d4-mk657_05fab4cb-323e-46a4-9b31-7e8f529e18cb/barbican-api/0.log" Oct 02 11:43:15 crc kubenswrapper[4771]: I1002 11:43:15.253687 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c4746c7d4-mk657_05fab4cb-323e-46a4-9b31-7e8f529e18cb/barbican-api-log/0.log" Oct 02 11:43:15 crc kubenswrapper[4771]: I1002 11:43:15.404486 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6998cd89db-l2xv7_b69ba24b-3361-44d2-8774-64230ab7a376/barbican-keystone-listener/0.log" Oct 02 11:43:15 crc kubenswrapper[4771]: I1002 11:43:15.552506 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6998cd89db-l2xv7_b69ba24b-3361-44d2-8774-64230ab7a376/barbican-keystone-listener-log/0.log" Oct 02 11:43:15 crc kubenswrapper[4771]: I1002 11:43:15.695810 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64df995667-zkf22_3439cba5-f86c-48f3-8d0b-587c02148975/barbican-worker/0.log" Oct 02 11:43:15 crc kubenswrapper[4771]: I1002 11:43:15.816894 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64df995667-zkf22_3439cba5-f86c-48f3-8d0b-587c02148975/barbican-worker-log/0.log" Oct 02 11:43:15 crc kubenswrapper[4771]: I1002 11:43:15.984467 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-w9zb4_4b64be94-5fae-49bf-83fd-b131c2d8aa24/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:16 crc kubenswrapper[4771]: I1002 11:43:16.281053 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_53db4baf-fd23-4d39-a4c8-0019beae7703/ceilometer-central-agent/0.log" Oct 02 11:43:16 crc kubenswrapper[4771]: I1002 11:43:16.356244 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_53db4baf-fd23-4d39-a4c8-0019beae7703/ceilometer-notification-agent/0.log" Oct 02 11:43:16 crc kubenswrapper[4771]: I1002 11:43:16.468732 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_53db4baf-fd23-4d39-a4c8-0019beae7703/proxy-httpd/0.log" Oct 02 11:43:16 crc kubenswrapper[4771]: I1002 11:43:16.578637 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_53db4baf-fd23-4d39-a4c8-0019beae7703/sg-core/0.log" Oct 02 11:43:16 crc kubenswrapper[4771]: I1002 11:43:16.847860 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_55b3670c-72d0-4a3d-aa68-f0b69f9653e9/cinder-api-log/0.log" Oct 02 11:43:16 crc kubenswrapper[4771]: I1002 11:43:16.855452 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_55b3670c-72d0-4a3d-aa68-f0b69f9653e9/cinder-api/0.log" Oct 02 11:43:17 crc kubenswrapper[4771]: I1002 11:43:17.114001 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61/cinder-scheduler/0.log" Oct 02 11:43:17 crc kubenswrapper[4771]: I1002 11:43:17.224673 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7e1d5ebb-a7e6-4509-bd86-f2fa34df2b61/probe/0.log" Oct 02 11:43:17 crc kubenswrapper[4771]: I1002 11:43:17.432380 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-xw9t8_9d368299-6a4f-4493-8460-3fc0e53f861c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:17 crc kubenswrapper[4771]: I1002 11:43:17.612241 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-dmdtw_0842954b-9245-4437-be8d-9cae5cda66c4/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:17 crc kubenswrapper[4771]: I1002 11:43:17.869509 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-ws9t2_5a1795e1-9162-428a-8ebe-fcc5273eed05/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:17 crc kubenswrapper[4771]: I1002 11:43:17.953562 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-vzhpr_faa0b798-d3ee-4469-b60b-15c77ecf1753/init/0.log" Oct 02 11:43:18 crc kubenswrapper[4771]: I1002 11:43:18.239686 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-vzhpr_faa0b798-d3ee-4469-b60b-15c77ecf1753/init/0.log" Oct 02 11:43:18 crc kubenswrapper[4771]: I1002 11:43:18.350849 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-vzhpr_faa0b798-d3ee-4469-b60b-15c77ecf1753/dnsmasq-dns/0.log" Oct 02 11:43:18 crc kubenswrapper[4771]: I1002 11:43:18.360760 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-qnj9r_7dbe44dc-bc16-4ba9-9975-4fd6b482ebd2/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:18 crc kubenswrapper[4771]: I1002 11:43:18.610165 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d6c1ce83-6b12-4748-a54e-e91b4e47eb11/glance-httpd/0.log" Oct 02 11:43:18 crc kubenswrapper[4771]: I1002 11:43:18.718701 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d6c1ce83-6b12-4748-a54e-e91b4e47eb11/glance-log/0.log" Oct 02 11:43:18 crc kubenswrapper[4771]: I1002 11:43:18.946600 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_606540a4-01ff-4915-89b2-f2a72c2fa922/glance-httpd/0.log" Oct 02 11:43:18 crc kubenswrapper[4771]: I1002 11:43:18.975327 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_606540a4-01ff-4915-89b2-f2a72c2fa922/glance-log/0.log" Oct 02 11:43:19 crc kubenswrapper[4771]: I1002 11:43:19.932741 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-9kh79_b2cab357-e81c-45c9-b0dc-6bdf44dd6092/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:20 crc kubenswrapper[4771]: I1002 11:43:20.116494 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-f9d48f8dd-cs4p5_3350ec2f-e01d-437e-99fd-49a97abfd448/heat-api/0.log" Oct 02 11:43:20 crc kubenswrapper[4771]: I1002 11:43:20.390325 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hbq9v_224fd5aa-8697-424f-8960-dcf3a2462981/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:20 crc kubenswrapper[4771]: I1002 11:43:20.549007 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-58b9db67cd-ctlwn_d8ca2772-c933-4f37-9f65-65ff09edfc8d/heat-cfnapi/0.log" Oct 02 11:43:20 crc kubenswrapper[4771]: I1002 11:43:20.811359 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29323321-8rm5d_c8986807-60cd-43ec-83f3-f9339b28c16f/keystone-cron/0.log" Oct 02 11:43:21 crc kubenswrapper[4771]: I1002 11:43:21.141994 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29323381-2gw54_6c79e9b7-b296-47b5-a27b-e72cbf77adf8/keystone-cron/0.log" Oct 02 11:43:21 crc kubenswrapper[4771]: I1002 11:43:21.170717 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-758b4b5558-442kz_883df2ca-c841-4882-b375-c74f7bf108f1/keystone-api/0.log" Oct 02 11:43:21 crc kubenswrapper[4771]: I1002 11:43:21.412228 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b20929e5-d840-461a-be05-c9f2783540f5/kube-state-metrics/0.log" Oct 02 11:43:21 crc kubenswrapper[4771]: I1002 11:43:21.796982 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-pczb2_db35f8f2-75a6-4c65-a44b-09230c27f953/logging-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:21 crc kubenswrapper[4771]: I1002 11:43:21.845341 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-zmv2x_7c7c8501-8ae8-47d5-b947-8b9ac02be829/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:22 crc kubenswrapper[4771]: I1002 11:43:22.444183 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_f57f79e3-d1a8-4830-a50b-6d414f041ec4/mysqld-exporter/0.log" Oct 02 11:43:22 crc kubenswrapper[4771]: I1002 11:43:22.805568 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-66d599bfb9-7qmxw_fb05af7b-d804-4adf-bb0f-dedaa49a7cd0/neutron-api/0.log" Oct 02 11:43:23 crc kubenswrapper[4771]: I1002 11:43:23.008227 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-66d599bfb9-7qmxw_fb05af7b-d804-4adf-bb0f-dedaa49a7cd0/neutron-httpd/0.log" Oct 02 11:43:23 crc kubenswrapper[4771]: I1002 11:43:23.305225 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-gx9cg_6697b08a-190c-48d2-9528-4cccd4efb520/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:24 crc kubenswrapper[4771]: I1002 11:43:24.249552 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_40e3003d-6fb9-455d-b3d6-0bebf099b5d1/nova-api-log/0.log" Oct 02 11:43:25 crc kubenswrapper[4771]: I1002 11:43:25.134843 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_40e3003d-6fb9-455d-b3d6-0bebf099b5d1/nova-api-api/0.log" Oct 02 11:43:25 crc kubenswrapper[4771]: I1002 11:43:25.136456 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_61c5ccc6-c403-431b-888a-71c9d79c702a/nova-cell0-conductor-conductor/0.log" Oct 02 11:43:25 crc kubenswrapper[4771]: I1002 11:43:25.530331 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ef7b8e9e-ce03-4409-986d-fa097341328c/nova-cell1-conductor-conductor/0.log" Oct 02 11:43:25 crc kubenswrapper[4771]: I1002 11:43:25.969921 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c8c49be1-946f-42eb-9767-6bb44807fb7d/nova-cell1-novncproxy-novncproxy/0.log" Oct 02 11:43:26 crc kubenswrapper[4771]: I1002 11:43:26.314480 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-fq28c_5242b9a6-8162-4a25-8821-5a8406cd2d86/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:26 crc kubenswrapper[4771]: I1002 11:43:26.654810 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_87c2141e-6fa2-45fc-9379-9a43a9cc9e06/nova-metadata-log/0.log" Oct 02 11:43:27 crc kubenswrapper[4771]: I1002 11:43:27.370540 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8d385667-ed65-45c5-8d75-b299520569b5/nova-scheduler-scheduler/0.log" Oct 02 11:43:27 crc kubenswrapper[4771]: I1002 11:43:27.892659 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_36a32369-c748-4356-be1b-5008d4f9bbe3/mysql-bootstrap/0.log" Oct 02 11:43:28 crc kubenswrapper[4771]: I1002 11:43:28.215511 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_36a32369-c748-4356-be1b-5008d4f9bbe3/mysql-bootstrap/0.log" Oct 02 11:43:28 crc kubenswrapper[4771]: I1002 11:43:28.540016 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_36a32369-c748-4356-be1b-5008d4f9bbe3/galera/0.log" Oct 02 11:43:29 crc kubenswrapper[4771]: I1002 11:43:29.035737 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_774a9026-9330-4c75-a2ca-33d5bccaf388/mysql-bootstrap/0.log" Oct 02 11:43:29 crc kubenswrapper[4771]: I1002 11:43:29.261803 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_774a9026-9330-4c75-a2ca-33d5bccaf388/mysql-bootstrap/0.log" Oct 02 11:43:29 crc kubenswrapper[4771]: I1002 11:43:29.500678 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_774a9026-9330-4c75-a2ca-33d5bccaf388/galera/0.log" Oct 02 11:43:30 crc kubenswrapper[4771]: I1002 11:43:30.093758 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_91134808-2774-4b54-8a59-09e9447dd87f/openstackclient/0.log" Oct 02 11:43:30 crc kubenswrapper[4771]: I1002 11:43:30.605272 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-5jf58_3388e290-ed26-4869-b6ce-c93670e7d9cf/ovn-controller/0.log" Oct 02 11:43:30 crc kubenswrapper[4771]: I1002 11:43:30.986098 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_87c2141e-6fa2-45fc-9379-9a43a9cc9e06/nova-metadata-metadata/0.log" Oct 02 11:43:31 crc kubenswrapper[4771]: I1002 11:43:31.042551 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-x9dc5_7e2e2fc3-4cf2-4455-a687-c07b61aed147/openstack-network-exporter/0.log" Oct 02 11:43:31 crc kubenswrapper[4771]: I1002 11:43:31.380033 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9qhq9_f3a8d6f9-c114-49ea-af29-8399e6a598cc/ovsdb-server-init/0.log" Oct 02 11:43:31 crc kubenswrapper[4771]: I1002 11:43:31.483757 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9qhq9_f3a8d6f9-c114-49ea-af29-8399e6a598cc/ovsdb-server-init/0.log" Oct 02 11:43:31 crc kubenswrapper[4771]: I1002 11:43:31.631676 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9qhq9_f3a8d6f9-c114-49ea-af29-8399e6a598cc/ovs-vswitchd/0.log" Oct 02 11:43:31 crc kubenswrapper[4771]: I1002 11:43:31.781286 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9qhq9_f3a8d6f9-c114-49ea-af29-8399e6a598cc/ovsdb-server/0.log" Oct 02 11:43:32 crc kubenswrapper[4771]: I1002 11:43:32.092643 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-ctttw_fb7648a0-0f4d-42ea-a800-1e2eaec5ec3d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:32 crc kubenswrapper[4771]: I1002 11:43:32.319412 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_98a17c5b-f1e3-476a-9728-5441e1182f50/openstack-network-exporter/0.log" Oct 02 11:43:32 crc kubenswrapper[4771]: I1002 11:43:32.468786 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_98a17c5b-f1e3-476a-9728-5441e1182f50/ovn-northd/0.log" Oct 02 11:43:32 crc kubenswrapper[4771]: I1002 11:43:32.684399 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cfd7c23d-a847-44c8-8e36-01ed1520831d/openstack-network-exporter/0.log" Oct 02 11:43:32 crc kubenswrapper[4771]: I1002 11:43:32.761044 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cfd7c23d-a847-44c8-8e36-01ed1520831d/ovsdbserver-nb/0.log" Oct 02 11:43:33 crc kubenswrapper[4771]: I1002 11:43:33.033798 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_795a79a7-e29c-496f-87c0-0bc8a1ef1570/openstack-network-exporter/0.log" Oct 02 11:43:33 crc kubenswrapper[4771]: I1002 11:43:33.259588 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_795a79a7-e29c-496f-87c0-0bc8a1ef1570/ovsdbserver-sb/0.log" Oct 02 11:43:33 crc kubenswrapper[4771]: I1002 11:43:33.691935 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-58587f9544-7mwm6_bd5227df-e476-4c2d-9ec2-694ab19dcb78/placement-api/0.log" Oct 02 11:43:33 crc kubenswrapper[4771]: I1002 11:43:33.777609 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-58587f9544-7mwm6_bd5227df-e476-4c2d-9ec2-694ab19dcb78/placement-log/0.log" Oct 02 11:43:34 crc kubenswrapper[4771]: I1002 11:43:34.000310 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6caafe76-3bdf-423b-abb6-c2057f835d38/init-config-reloader/0.log" Oct 02 11:43:34 crc kubenswrapper[4771]: I1002 11:43:34.222830 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6caafe76-3bdf-423b-abb6-c2057f835d38/init-config-reloader/0.log" Oct 02 11:43:34 crc kubenswrapper[4771]: I1002 11:43:34.301544 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6caafe76-3bdf-423b-abb6-c2057f835d38/config-reloader/0.log" Oct 02 11:43:34 crc kubenswrapper[4771]: I1002 11:43:34.406840 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6caafe76-3bdf-423b-abb6-c2057f835d38/prometheus/0.log" Oct 02 11:43:34 crc kubenswrapper[4771]: I1002 11:43:34.495630 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6caafe76-3bdf-423b-abb6-c2057f835d38/thanos-sidecar/0.log" Oct 02 11:43:34 crc kubenswrapper[4771]: I1002 11:43:34.750468 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3fa19dbc-a10c-4863-bf25-035f9d6579e9/setup-container/0.log" Oct 02 11:43:35 crc kubenswrapper[4771]: I1002 11:43:35.001577 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3fa19dbc-a10c-4863-bf25-035f9d6579e9/setup-container/0.log" Oct 02 11:43:35 crc kubenswrapper[4771]: I1002 11:43:35.054741 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3fa19dbc-a10c-4863-bf25-035f9d6579e9/rabbitmq/0.log" Oct 02 11:43:35 crc kubenswrapper[4771]: I1002 11:43:35.278285 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_07281b0c-ff0a-4195-9ddb-521645890aee/setup-container/0.log" Oct 02 11:43:35 crc kubenswrapper[4771]: I1002 11:43:35.534549 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_07281b0c-ff0a-4195-9ddb-521645890aee/setup-container/0.log" Oct 02 11:43:36 crc kubenswrapper[4771]: I1002 11:43:36.025596 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-5n2v5_b22f05d8-7526-4e04-bac0-cdd965cac0cf/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:36 crc kubenswrapper[4771]: I1002 11:43:36.434110 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-54b646f5dd-mdqrv_2f645324-58c1-45d2-9859-aa05f9644975/heat-engine/0.log" Oct 02 11:43:36 crc kubenswrapper[4771]: I1002 11:43:36.496771 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-fbffz_e7004c3e-7b27-47fc-accf-7156c0fe77b0/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:36 crc kubenswrapper[4771]: I1002 11:43:36.763294 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-km5kj_58b8222a-a342-4b38-8aa6-47c6ea065493/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:37 crc kubenswrapper[4771]: I1002 11:43:37.049500 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-dsv2t_2cb23df5-8bb9-41fa-999e-6a563575b760/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:37 crc kubenswrapper[4771]: I1002 11:43:37.331316 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-5g9vt_2315a9df-8051-4ec6-98a2-d8e0e32f598b/ssh-known-hosts-edpm-deployment/0.log" Oct 02 11:43:37 crc kubenswrapper[4771]: I1002 11:43:37.838390 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-668c9ddd79-9f298_285035a2-e600-4571-9462-e029a84b3779/proxy-server/0.log" Oct 02 11:43:37 crc kubenswrapper[4771]: I1002 11:43:37.973189 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-668c9ddd79-9f298_285035a2-e600-4571-9462-e029a84b3779/proxy-httpd/0.log" Oct 02 11:43:38 crc kubenswrapper[4771]: I1002 11:43:38.198890 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-c9bwn_087b7d39-d0cf-4896-8220-d79e94810ff5/swift-ring-rebalance/0.log" Oct 02 11:43:38 crc kubenswrapper[4771]: I1002 11:43:38.468491 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/account-auditor/0.log" Oct 02 11:43:38 crc kubenswrapper[4771]: I1002 11:43:38.582928 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/account-reaper/0.log" Oct 02 11:43:38 crc kubenswrapper[4771]: I1002 11:43:38.778736 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/account-server/0.log" Oct 02 11:43:38 crc kubenswrapper[4771]: I1002 11:43:38.833276 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/account-replicator/0.log" Oct 02 11:43:39 crc kubenswrapper[4771]: I1002 11:43:39.006235 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/container-auditor/0.log" Oct 02 11:43:39 crc kubenswrapper[4771]: I1002 11:43:39.121173 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/container-replicator/0.log" Oct 02 11:43:39 crc kubenswrapper[4771]: I1002 11:43:39.209230 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/container-server/0.log" Oct 02 11:43:39 crc kubenswrapper[4771]: I1002 11:43:39.326388 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/container-updater/0.log" Oct 02 11:43:39 crc kubenswrapper[4771]: I1002 11:43:39.458642 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/object-auditor/0.log" Oct 02 11:43:39 crc kubenswrapper[4771]: I1002 11:43:39.633068 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/object-expirer/0.log" Oct 02 11:43:39 crc kubenswrapper[4771]: I1002 11:43:39.735780 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_07281b0c-ff0a-4195-9ddb-521645890aee/rabbitmq/0.log" Oct 02 11:43:39 crc kubenswrapper[4771]: I1002 11:43:39.740505 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/object-replicator/0.log" Oct 02 11:43:39 crc kubenswrapper[4771]: I1002 11:43:39.830956 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/object-server/0.log" Oct 02 11:43:39 crc kubenswrapper[4771]: I1002 11:43:39.967060 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/rsync/0.log" Oct 02 11:43:39 crc kubenswrapper[4771]: I1002 11:43:39.972637 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/object-updater/0.log" Oct 02 11:43:40 crc kubenswrapper[4771]: I1002 11:43:40.117651 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4c4deacb-7b2f-49d1-8c7e-b87913f027d1/swift-recon-cron/0.log" Oct 02 11:43:40 crc kubenswrapper[4771]: I1002 11:43:40.308007 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-qhxvk_21219956-7fb8-4a13-8b59-6f55702fe548/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:40 crc kubenswrapper[4771]: I1002 11:43:40.667247 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-j9qs6_d6e37706-8393-4c3c-9941-3e615b320c51/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:40 crc kubenswrapper[4771]: I1002 11:43:40.731683 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_57b7360c-088e-4c95-a10b-e7d034d2f83f/tempest-tests-tempest-tests-runner/0.log" Oct 02 11:43:40 crc kubenswrapper[4771]: I1002 11:43:40.899083 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_3e55f1c1-c148-4ee4-80fb-29c6029f2d2d/test-operator-logs-container/0.log" Oct 02 11:43:41 crc kubenswrapper[4771]: I1002 11:43:41.126219 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-cdkdq_7b873f5d-5502-4c9b-ba58-9eba9b62739b/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 11:43:42 crc kubenswrapper[4771]: I1002 11:43:42.145850 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:43:42 crc kubenswrapper[4771]: I1002 11:43:42.145907 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:43:42 crc kubenswrapper[4771]: I1002 11:43:42.145967 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 11:43:42 crc kubenswrapper[4771]: I1002 11:43:42.149902 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3bb74c9f46be75efdbfc800b6d2f38f7da79e737ae4bf6020f5528c548db4dc6"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:43:42 crc kubenswrapper[4771]: I1002 11:43:42.149986 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://3bb74c9f46be75efdbfc800b6d2f38f7da79e737ae4bf6020f5528c548db4dc6" gracePeriod=600 Oct 02 11:43:43 crc kubenswrapper[4771]: I1002 11:43:43.088523 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="3bb74c9f46be75efdbfc800b6d2f38f7da79e737ae4bf6020f5528c548db4dc6" exitCode=0 Oct 02 11:43:43 crc kubenswrapper[4771]: I1002 11:43:43.089112 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"3bb74c9f46be75efdbfc800b6d2f38f7da79e737ae4bf6020f5528c548db4dc6"} Oct 02 11:43:43 crc kubenswrapper[4771]: I1002 11:43:43.089161 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5"} Oct 02 11:43:43 crc kubenswrapper[4771]: I1002 11:43:43.089182 4771 scope.go:117] "RemoveContainer" containerID="8cbe5d2a73e71064329aee71c315af953434b4ef10e8d34dc68f6b0ddd2034f9" Oct 02 11:43:56 crc kubenswrapper[4771]: I1002 11:43:56.885087 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_785673c9-400c-474f-915d-30e8da38b311/memcached/0.log" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.341361 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zns7f"] Oct 02 11:44:19 crc kubenswrapper[4771]: E1002 11:44:19.343231 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" containerName="extract-content" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.343340 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" containerName="extract-content" Oct 02 11:44:19 crc kubenswrapper[4771]: E1002 11:44:19.343384 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" containerName="extract-utilities" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.343392 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" containerName="extract-utilities" Oct 02 11:44:19 crc kubenswrapper[4771]: E1002 11:44:19.343406 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" containerName="registry-server" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.343412 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" containerName="registry-server" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.345643 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b9ea4ef-3ad9-4924-a325-036c9d6dd7fe" containerName="registry-server" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.348121 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.422669 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zns7f"] Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.438680 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cc680e7-1fed-4d99-b65e-37737d08a38b-catalog-content\") pod \"redhat-marketplace-zns7f\" (UID: \"4cc680e7-1fed-4d99-b65e-37737d08a38b\") " pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.438738 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brgpg\" (UniqueName: \"kubernetes.io/projected/4cc680e7-1fed-4d99-b65e-37737d08a38b-kube-api-access-brgpg\") pod \"redhat-marketplace-zns7f\" (UID: \"4cc680e7-1fed-4d99-b65e-37737d08a38b\") " pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.438779 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cc680e7-1fed-4d99-b65e-37737d08a38b-utilities\") pod \"redhat-marketplace-zns7f\" (UID: \"4cc680e7-1fed-4d99-b65e-37737d08a38b\") " pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.540772 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cc680e7-1fed-4d99-b65e-37737d08a38b-catalog-content\") pod \"redhat-marketplace-zns7f\" (UID: \"4cc680e7-1fed-4d99-b65e-37737d08a38b\") " pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.540819 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brgpg\" (UniqueName: \"kubernetes.io/projected/4cc680e7-1fed-4d99-b65e-37737d08a38b-kube-api-access-brgpg\") pod \"redhat-marketplace-zns7f\" (UID: \"4cc680e7-1fed-4d99-b65e-37737d08a38b\") " pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.540849 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cc680e7-1fed-4d99-b65e-37737d08a38b-utilities\") pod \"redhat-marketplace-zns7f\" (UID: \"4cc680e7-1fed-4d99-b65e-37737d08a38b\") " pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.542205 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cc680e7-1fed-4d99-b65e-37737d08a38b-utilities\") pod \"redhat-marketplace-zns7f\" (UID: \"4cc680e7-1fed-4d99-b65e-37737d08a38b\") " pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.542231 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cc680e7-1fed-4d99-b65e-37737d08a38b-catalog-content\") pod \"redhat-marketplace-zns7f\" (UID: \"4cc680e7-1fed-4d99-b65e-37737d08a38b\") " pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.592968 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brgpg\" (UniqueName: \"kubernetes.io/projected/4cc680e7-1fed-4d99-b65e-37737d08a38b-kube-api-access-brgpg\") pod \"redhat-marketplace-zns7f\" (UID: \"4cc680e7-1fed-4d99-b65e-37737d08a38b\") " pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:19 crc kubenswrapper[4771]: I1002 11:44:19.670097 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:20 crc kubenswrapper[4771]: I1002 11:44:20.639856 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zns7f"] Oct 02 11:44:21 crc kubenswrapper[4771]: I1002 11:44:21.537105 4771 generic.go:334] "Generic (PLEG): container finished" podID="4cc680e7-1fed-4d99-b65e-37737d08a38b" containerID="3dd05234fbccaa46d6b5ff95eeb22ae4b56b94f36793789c11ab612025177c63" exitCode=0 Oct 02 11:44:21 crc kubenswrapper[4771]: I1002 11:44:21.537199 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zns7f" event={"ID":"4cc680e7-1fed-4d99-b65e-37737d08a38b","Type":"ContainerDied","Data":"3dd05234fbccaa46d6b5ff95eeb22ae4b56b94f36793789c11ab612025177c63"} Oct 02 11:44:21 crc kubenswrapper[4771]: I1002 11:44:21.537514 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zns7f" event={"ID":"4cc680e7-1fed-4d99-b65e-37737d08a38b","Type":"ContainerStarted","Data":"99c577ec2ada0f81f4187d523f003e5894fc8f09282a5805ecf3fd6e963e24ce"} Oct 02 11:44:21 crc kubenswrapper[4771]: I1002 11:44:21.542768 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:44:23 crc kubenswrapper[4771]: I1002 11:44:23.575297 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zns7f" event={"ID":"4cc680e7-1fed-4d99-b65e-37737d08a38b","Type":"ContainerStarted","Data":"77acf6909ace533da26b352108b5d6f3218c9c67d7c8cc33a525de498bc544f7"} Oct 02 11:44:24 crc kubenswrapper[4771]: I1002 11:44:24.587615 4771 generic.go:334] "Generic (PLEG): container finished" podID="4cc680e7-1fed-4d99-b65e-37737d08a38b" containerID="77acf6909ace533da26b352108b5d6f3218c9c67d7c8cc33a525de498bc544f7" exitCode=0 Oct 02 11:44:24 crc kubenswrapper[4771]: I1002 11:44:24.587686 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zns7f" event={"ID":"4cc680e7-1fed-4d99-b65e-37737d08a38b","Type":"ContainerDied","Data":"77acf6909ace533da26b352108b5d6f3218c9c67d7c8cc33a525de498bc544f7"} Oct 02 11:44:25 crc kubenswrapper[4771]: I1002 11:44:25.605727 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zns7f" event={"ID":"4cc680e7-1fed-4d99-b65e-37737d08a38b","Type":"ContainerStarted","Data":"751770e9a40f1f34af71060331d732225d1e75cf1362b91a2595348fcae1af75"} Oct 02 11:44:25 crc kubenswrapper[4771]: I1002 11:44:25.638256 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zns7f" podStartSLOduration=2.826298505 podStartE2EDuration="6.638211962s" podCreationTimestamp="2025-10-02 11:44:19 +0000 UTC" firstStartedPulling="2025-10-02 11:44:21.53943574 +0000 UTC m=+7649.187120807" lastFinishedPulling="2025-10-02 11:44:25.351349197 +0000 UTC m=+7652.999034264" observedRunningTime="2025-10-02 11:44:25.630527404 +0000 UTC m=+7653.278212471" watchObservedRunningTime="2025-10-02 11:44:25.638211962 +0000 UTC m=+7653.285897029" Oct 02 11:44:29 crc kubenswrapper[4771]: I1002 11:44:29.670856 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:29 crc kubenswrapper[4771]: I1002 11:44:29.671429 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:29 crc kubenswrapper[4771]: I1002 11:44:29.725200 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:30 crc kubenswrapper[4771]: I1002 11:44:30.748339 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:30 crc kubenswrapper[4771]: I1002 11:44:30.844470 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zns7f"] Oct 02 11:44:32 crc kubenswrapper[4771]: I1002 11:44:32.693308 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zns7f" podUID="4cc680e7-1fed-4d99-b65e-37737d08a38b" containerName="registry-server" containerID="cri-o://751770e9a40f1f34af71060331d732225d1e75cf1362b91a2595348fcae1af75" gracePeriod=2 Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.251326 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.424321 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brgpg\" (UniqueName: \"kubernetes.io/projected/4cc680e7-1fed-4d99-b65e-37737d08a38b-kube-api-access-brgpg\") pod \"4cc680e7-1fed-4d99-b65e-37737d08a38b\" (UID: \"4cc680e7-1fed-4d99-b65e-37737d08a38b\") " Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.424832 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cc680e7-1fed-4d99-b65e-37737d08a38b-utilities\") pod \"4cc680e7-1fed-4d99-b65e-37737d08a38b\" (UID: \"4cc680e7-1fed-4d99-b65e-37737d08a38b\") " Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.425040 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cc680e7-1fed-4d99-b65e-37737d08a38b-catalog-content\") pod \"4cc680e7-1fed-4d99-b65e-37737d08a38b\" (UID: \"4cc680e7-1fed-4d99-b65e-37737d08a38b\") " Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.425899 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cc680e7-1fed-4d99-b65e-37737d08a38b-utilities" (OuterVolumeSpecName: "utilities") pod "4cc680e7-1fed-4d99-b65e-37737d08a38b" (UID: "4cc680e7-1fed-4d99-b65e-37737d08a38b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.431670 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cc680e7-1fed-4d99-b65e-37737d08a38b-kube-api-access-brgpg" (OuterVolumeSpecName: "kube-api-access-brgpg") pod "4cc680e7-1fed-4d99-b65e-37737d08a38b" (UID: "4cc680e7-1fed-4d99-b65e-37737d08a38b"). InnerVolumeSpecName "kube-api-access-brgpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.438970 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cc680e7-1fed-4d99-b65e-37737d08a38b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4cc680e7-1fed-4d99-b65e-37737d08a38b" (UID: "4cc680e7-1fed-4d99-b65e-37737d08a38b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.528422 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cc680e7-1fed-4d99-b65e-37737d08a38b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.528472 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brgpg\" (UniqueName: \"kubernetes.io/projected/4cc680e7-1fed-4d99-b65e-37737d08a38b-kube-api-access-brgpg\") on node \"crc\" DevicePath \"\"" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.528487 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cc680e7-1fed-4d99-b65e-37737d08a38b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.719220 4771 generic.go:334] "Generic (PLEG): container finished" podID="4cc680e7-1fed-4d99-b65e-37737d08a38b" containerID="751770e9a40f1f34af71060331d732225d1e75cf1362b91a2595348fcae1af75" exitCode=0 Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.719265 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zns7f" event={"ID":"4cc680e7-1fed-4d99-b65e-37737d08a38b","Type":"ContainerDied","Data":"751770e9a40f1f34af71060331d732225d1e75cf1362b91a2595348fcae1af75"} Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.719270 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zns7f" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.719291 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zns7f" event={"ID":"4cc680e7-1fed-4d99-b65e-37737d08a38b","Type":"ContainerDied","Data":"99c577ec2ada0f81f4187d523f003e5894fc8f09282a5805ecf3fd6e963e24ce"} Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.719326 4771 scope.go:117] "RemoveContainer" containerID="751770e9a40f1f34af71060331d732225d1e75cf1362b91a2595348fcae1af75" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.751198 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zns7f"] Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.754476 4771 scope.go:117] "RemoveContainer" containerID="77acf6909ace533da26b352108b5d6f3218c9c67d7c8cc33a525de498bc544f7" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.762480 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zns7f"] Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.792886 4771 scope.go:117] "RemoveContainer" containerID="3dd05234fbccaa46d6b5ff95eeb22ae4b56b94f36793789c11ab612025177c63" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.835014 4771 scope.go:117] "RemoveContainer" containerID="751770e9a40f1f34af71060331d732225d1e75cf1362b91a2595348fcae1af75" Oct 02 11:44:33 crc kubenswrapper[4771]: E1002 11:44:33.838305 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"751770e9a40f1f34af71060331d732225d1e75cf1362b91a2595348fcae1af75\": container with ID starting with 751770e9a40f1f34af71060331d732225d1e75cf1362b91a2595348fcae1af75 not found: ID does not exist" containerID="751770e9a40f1f34af71060331d732225d1e75cf1362b91a2595348fcae1af75" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.838702 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"751770e9a40f1f34af71060331d732225d1e75cf1362b91a2595348fcae1af75"} err="failed to get container status \"751770e9a40f1f34af71060331d732225d1e75cf1362b91a2595348fcae1af75\": rpc error: code = NotFound desc = could not find container \"751770e9a40f1f34af71060331d732225d1e75cf1362b91a2595348fcae1af75\": container with ID starting with 751770e9a40f1f34af71060331d732225d1e75cf1362b91a2595348fcae1af75 not found: ID does not exist" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.838775 4771 scope.go:117] "RemoveContainer" containerID="77acf6909ace533da26b352108b5d6f3218c9c67d7c8cc33a525de498bc544f7" Oct 02 11:44:33 crc kubenswrapper[4771]: E1002 11:44:33.840397 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77acf6909ace533da26b352108b5d6f3218c9c67d7c8cc33a525de498bc544f7\": container with ID starting with 77acf6909ace533da26b352108b5d6f3218c9c67d7c8cc33a525de498bc544f7 not found: ID does not exist" containerID="77acf6909ace533da26b352108b5d6f3218c9c67d7c8cc33a525de498bc544f7" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.840462 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77acf6909ace533da26b352108b5d6f3218c9c67d7c8cc33a525de498bc544f7"} err="failed to get container status \"77acf6909ace533da26b352108b5d6f3218c9c67d7c8cc33a525de498bc544f7\": rpc error: code = NotFound desc = could not find container \"77acf6909ace533da26b352108b5d6f3218c9c67d7c8cc33a525de498bc544f7\": container with ID starting with 77acf6909ace533da26b352108b5d6f3218c9c67d7c8cc33a525de498bc544f7 not found: ID does not exist" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.840533 4771 scope.go:117] "RemoveContainer" containerID="3dd05234fbccaa46d6b5ff95eeb22ae4b56b94f36793789c11ab612025177c63" Oct 02 11:44:33 crc kubenswrapper[4771]: E1002 11:44:33.841513 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dd05234fbccaa46d6b5ff95eeb22ae4b56b94f36793789c11ab612025177c63\": container with ID starting with 3dd05234fbccaa46d6b5ff95eeb22ae4b56b94f36793789c11ab612025177c63 not found: ID does not exist" containerID="3dd05234fbccaa46d6b5ff95eeb22ae4b56b94f36793789c11ab612025177c63" Oct 02 11:44:33 crc kubenswrapper[4771]: I1002 11:44:33.841579 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dd05234fbccaa46d6b5ff95eeb22ae4b56b94f36793789c11ab612025177c63"} err="failed to get container status \"3dd05234fbccaa46d6b5ff95eeb22ae4b56b94f36793789c11ab612025177c63\": rpc error: code = NotFound desc = could not find container \"3dd05234fbccaa46d6b5ff95eeb22ae4b56b94f36793789c11ab612025177c63\": container with ID starting with 3dd05234fbccaa46d6b5ff95eeb22ae4b56b94f36793789c11ab612025177c63 not found: ID does not exist" Oct 02 11:44:35 crc kubenswrapper[4771]: I1002 11:44:35.699152 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cc680e7-1fed-4d99-b65e-37737d08a38b" path="/var/lib/kubelet/pods/4cc680e7-1fed-4d99-b65e-37737d08a38b/volumes" Oct 02 11:44:51 crc kubenswrapper[4771]: I1002 11:44:51.935253 4771 generic.go:334] "Generic (PLEG): container finished" podID="85fb7ae5-5b81-4fc7-9527-ab0df063d173" containerID="400d3412c9eea1469ff0b878295f2fcb0f2b31f92d51dfed0f5b3936c8996a9e" exitCode=0 Oct 02 11:44:51 crc kubenswrapper[4771]: I1002 11:44:51.935338 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pc9ng/crc-debug-ks887" event={"ID":"85fb7ae5-5b81-4fc7-9527-ab0df063d173","Type":"ContainerDied","Data":"400d3412c9eea1469ff0b878295f2fcb0f2b31f92d51dfed0f5b3936c8996a9e"} Oct 02 11:44:53 crc kubenswrapper[4771]: I1002 11:44:53.082343 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/crc-debug-ks887" Oct 02 11:44:53 crc kubenswrapper[4771]: I1002 11:44:53.112285 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/85fb7ae5-5b81-4fc7-9527-ab0df063d173-host\") pod \"85fb7ae5-5b81-4fc7-9527-ab0df063d173\" (UID: \"85fb7ae5-5b81-4fc7-9527-ab0df063d173\") " Oct 02 11:44:53 crc kubenswrapper[4771]: I1002 11:44:53.112355 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p74c\" (UniqueName: \"kubernetes.io/projected/85fb7ae5-5b81-4fc7-9527-ab0df063d173-kube-api-access-9p74c\") pod \"85fb7ae5-5b81-4fc7-9527-ab0df063d173\" (UID: \"85fb7ae5-5b81-4fc7-9527-ab0df063d173\") " Oct 02 11:44:53 crc kubenswrapper[4771]: I1002 11:44:53.112416 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/85fb7ae5-5b81-4fc7-9527-ab0df063d173-host" (OuterVolumeSpecName: "host") pod "85fb7ae5-5b81-4fc7-9527-ab0df063d173" (UID: "85fb7ae5-5b81-4fc7-9527-ab0df063d173"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:44:53 crc kubenswrapper[4771]: I1002 11:44:53.113152 4771 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/85fb7ae5-5b81-4fc7-9527-ab0df063d173-host\") on node \"crc\" DevicePath \"\"" Oct 02 11:44:53 crc kubenswrapper[4771]: I1002 11:44:53.124722 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pc9ng/crc-debug-ks887"] Oct 02 11:44:53 crc kubenswrapper[4771]: I1002 11:44:53.133420 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85fb7ae5-5b81-4fc7-9527-ab0df063d173-kube-api-access-9p74c" (OuterVolumeSpecName: "kube-api-access-9p74c") pod "85fb7ae5-5b81-4fc7-9527-ab0df063d173" (UID: "85fb7ae5-5b81-4fc7-9527-ab0df063d173"). InnerVolumeSpecName "kube-api-access-9p74c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:44:53 crc kubenswrapper[4771]: I1002 11:44:53.136347 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pc9ng/crc-debug-ks887"] Oct 02 11:44:53 crc kubenswrapper[4771]: I1002 11:44:53.216189 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p74c\" (UniqueName: \"kubernetes.io/projected/85fb7ae5-5b81-4fc7-9527-ab0df063d173-kube-api-access-9p74c\") on node \"crc\" DevicePath \"\"" Oct 02 11:44:53 crc kubenswrapper[4771]: I1002 11:44:53.702805 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85fb7ae5-5b81-4fc7-9527-ab0df063d173" path="/var/lib/kubelet/pods/85fb7ae5-5b81-4fc7-9527-ab0df063d173/volumes" Oct 02 11:44:53 crc kubenswrapper[4771]: I1002 11:44:53.961746 4771 scope.go:117] "RemoveContainer" containerID="400d3412c9eea1469ff0b878295f2fcb0f2b31f92d51dfed0f5b3936c8996a9e" Oct 02 11:44:53 crc kubenswrapper[4771]: I1002 11:44:53.961839 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/crc-debug-ks887" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.329172 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pc9ng/crc-debug-d6lf7"] Oct 02 11:44:54 crc kubenswrapper[4771]: E1002 11:44:54.329977 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85fb7ae5-5b81-4fc7-9527-ab0df063d173" containerName="container-00" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.329999 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="85fb7ae5-5b81-4fc7-9527-ab0df063d173" containerName="container-00" Oct 02 11:44:54 crc kubenswrapper[4771]: E1002 11:44:54.330019 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc680e7-1fed-4d99-b65e-37737d08a38b" containerName="extract-utilities" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.330031 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc680e7-1fed-4d99-b65e-37737d08a38b" containerName="extract-utilities" Oct 02 11:44:54 crc kubenswrapper[4771]: E1002 11:44:54.330074 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc680e7-1fed-4d99-b65e-37737d08a38b" containerName="registry-server" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.330753 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc680e7-1fed-4d99-b65e-37737d08a38b" containerName="registry-server" Oct 02 11:44:54 crc kubenswrapper[4771]: E1002 11:44:54.330798 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc680e7-1fed-4d99-b65e-37737d08a38b" containerName="extract-content" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.330805 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc680e7-1fed-4d99-b65e-37737d08a38b" containerName="extract-content" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.331452 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="85fb7ae5-5b81-4fc7-9527-ab0df063d173" containerName="container-00" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.331473 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cc680e7-1fed-4d99-b65e-37737d08a38b" containerName="registry-server" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.333473 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.444179 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsqxn\" (UniqueName: \"kubernetes.io/projected/61229121-9b45-47d2-858c-5cc345408fc6-kube-api-access-wsqxn\") pod \"crc-debug-d6lf7\" (UID: \"61229121-9b45-47d2-858c-5cc345408fc6\") " pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.444700 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/61229121-9b45-47d2-858c-5cc345408fc6-host\") pod \"crc-debug-d6lf7\" (UID: \"61229121-9b45-47d2-858c-5cc345408fc6\") " pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.546775 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsqxn\" (UniqueName: \"kubernetes.io/projected/61229121-9b45-47d2-858c-5cc345408fc6-kube-api-access-wsqxn\") pod \"crc-debug-d6lf7\" (UID: \"61229121-9b45-47d2-858c-5cc345408fc6\") " pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.546914 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/61229121-9b45-47d2-858c-5cc345408fc6-host\") pod \"crc-debug-d6lf7\" (UID: \"61229121-9b45-47d2-858c-5cc345408fc6\") " pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.547031 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/61229121-9b45-47d2-858c-5cc345408fc6-host\") pod \"crc-debug-d6lf7\" (UID: \"61229121-9b45-47d2-858c-5cc345408fc6\") " pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.570647 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsqxn\" (UniqueName: \"kubernetes.io/projected/61229121-9b45-47d2-858c-5cc345408fc6-kube-api-access-wsqxn\") pod \"crc-debug-d6lf7\" (UID: \"61229121-9b45-47d2-858c-5cc345408fc6\") " pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.653084 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.980668 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" event={"ID":"61229121-9b45-47d2-858c-5cc345408fc6","Type":"ContainerStarted","Data":"3bb7ea0a141a602949a8bd1148c94c21c4548c1d184717a4655bd3bbcfb4afeb"} Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.981246 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" event={"ID":"61229121-9b45-47d2-858c-5cc345408fc6","Type":"ContainerStarted","Data":"d06b275ee9b6d43b4766b13f9906783991069b49ab0dae347cb7e4622090233b"} Oct 02 11:44:54 crc kubenswrapper[4771]: I1002 11:44:54.997801 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" podStartSLOduration=0.997768973 podStartE2EDuration="997.768973ms" podCreationTimestamp="2025-10-02 11:44:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:44:54.993494163 +0000 UTC m=+7682.641179230" watchObservedRunningTime="2025-10-02 11:44:54.997768973 +0000 UTC m=+7682.645454040" Oct 02 11:44:56 crc kubenswrapper[4771]: I1002 11:44:56.010694 4771 generic.go:334] "Generic (PLEG): container finished" podID="61229121-9b45-47d2-858c-5cc345408fc6" containerID="3bb7ea0a141a602949a8bd1148c94c21c4548c1d184717a4655bd3bbcfb4afeb" exitCode=0 Oct 02 11:44:56 crc kubenswrapper[4771]: I1002 11:44:56.010761 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" event={"ID":"61229121-9b45-47d2-858c-5cc345408fc6","Type":"ContainerDied","Data":"3bb7ea0a141a602949a8bd1148c94c21c4548c1d184717a4655bd3bbcfb4afeb"} Oct 02 11:44:57 crc kubenswrapper[4771]: I1002 11:44:57.158953 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" Oct 02 11:44:57 crc kubenswrapper[4771]: I1002 11:44:57.210325 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/61229121-9b45-47d2-858c-5cc345408fc6-host\") pod \"61229121-9b45-47d2-858c-5cc345408fc6\" (UID: \"61229121-9b45-47d2-858c-5cc345408fc6\") " Oct 02 11:44:57 crc kubenswrapper[4771]: I1002 11:44:57.210473 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsqxn\" (UniqueName: \"kubernetes.io/projected/61229121-9b45-47d2-858c-5cc345408fc6-kube-api-access-wsqxn\") pod \"61229121-9b45-47d2-858c-5cc345408fc6\" (UID: \"61229121-9b45-47d2-858c-5cc345408fc6\") " Oct 02 11:44:57 crc kubenswrapper[4771]: I1002 11:44:57.210500 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61229121-9b45-47d2-858c-5cc345408fc6-host" (OuterVolumeSpecName: "host") pod "61229121-9b45-47d2-858c-5cc345408fc6" (UID: "61229121-9b45-47d2-858c-5cc345408fc6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:44:57 crc kubenswrapper[4771]: I1002 11:44:57.211102 4771 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/61229121-9b45-47d2-858c-5cc345408fc6-host\") on node \"crc\" DevicePath \"\"" Oct 02 11:44:57 crc kubenswrapper[4771]: I1002 11:44:57.222496 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61229121-9b45-47d2-858c-5cc345408fc6-kube-api-access-wsqxn" (OuterVolumeSpecName: "kube-api-access-wsqxn") pod "61229121-9b45-47d2-858c-5cc345408fc6" (UID: "61229121-9b45-47d2-858c-5cc345408fc6"). InnerVolumeSpecName "kube-api-access-wsqxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:44:57 crc kubenswrapper[4771]: I1002 11:44:57.312766 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsqxn\" (UniqueName: \"kubernetes.io/projected/61229121-9b45-47d2-858c-5cc345408fc6-kube-api-access-wsqxn\") on node \"crc\" DevicePath \"\"" Oct 02 11:44:58 crc kubenswrapper[4771]: I1002 11:44:58.034521 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" event={"ID":"61229121-9b45-47d2-858c-5cc345408fc6","Type":"ContainerDied","Data":"d06b275ee9b6d43b4766b13f9906783991069b49ab0dae347cb7e4622090233b"} Oct 02 11:44:58 crc kubenswrapper[4771]: I1002 11:44:58.034851 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d06b275ee9b6d43b4766b13f9906783991069b49ab0dae347cb7e4622090233b" Oct 02 11:44:58 crc kubenswrapper[4771]: I1002 11:44:58.034591 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/crc-debug-d6lf7" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.243717 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd"] Oct 02 11:45:00 crc kubenswrapper[4771]: E1002 11:45:00.244713 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61229121-9b45-47d2-858c-5cc345408fc6" containerName="container-00" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.244735 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="61229121-9b45-47d2-858c-5cc345408fc6" containerName="container-00" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.245048 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="61229121-9b45-47d2-858c-5cc345408fc6" containerName="container-00" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.249363 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.256262 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd"] Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.268426 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.269803 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.281100 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-config-volume\") pod \"collect-profiles-29323425-dwdjd\" (UID: \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.281165 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-secret-volume\") pod \"collect-profiles-29323425-dwdjd\" (UID: \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.281214 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg8z6\" (UniqueName: \"kubernetes.io/projected/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-kube-api-access-lg8z6\") pod \"collect-profiles-29323425-dwdjd\" (UID: \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.383438 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-config-volume\") pod \"collect-profiles-29323425-dwdjd\" (UID: \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.383474 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-secret-volume\") pod \"collect-profiles-29323425-dwdjd\" (UID: \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.383548 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg8z6\" (UniqueName: \"kubernetes.io/projected/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-kube-api-access-lg8z6\") pod \"collect-profiles-29323425-dwdjd\" (UID: \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.384290 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-config-volume\") pod \"collect-profiles-29323425-dwdjd\" (UID: \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.408095 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-secret-volume\") pod \"collect-profiles-29323425-dwdjd\" (UID: \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.418939 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg8z6\" (UniqueName: \"kubernetes.io/projected/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-kube-api-access-lg8z6\") pod \"collect-profiles-29323425-dwdjd\" (UID: \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" Oct 02 11:45:00 crc kubenswrapper[4771]: I1002 11:45:00.576627 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" Oct 02 11:45:01 crc kubenswrapper[4771]: I1002 11:45:01.072801 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd"] Oct 02 11:45:02 crc kubenswrapper[4771]: I1002 11:45:02.080720 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" event={"ID":"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e","Type":"ContainerStarted","Data":"90b4d26e4d3364abc0a61b52c21177fee27abafd7f691f7e37a082efe8b96309"} Oct 02 11:45:02 crc kubenswrapper[4771]: I1002 11:45:02.081089 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" event={"ID":"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e","Type":"ContainerStarted","Data":"ab80705bb90f0935dce90738afa682ea196e80e2dc99e17624fffebc05a82f53"} Oct 02 11:45:02 crc kubenswrapper[4771]: I1002 11:45:02.109714 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" podStartSLOduration=2.109693595 podStartE2EDuration="2.109693595s" podCreationTimestamp="2025-10-02 11:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:45:02.096641217 +0000 UTC m=+7689.744326284" watchObservedRunningTime="2025-10-02 11:45:02.109693595 +0000 UTC m=+7689.757378652" Oct 02 11:45:03 crc kubenswrapper[4771]: I1002 11:45:03.100513 4771 generic.go:334] "Generic (PLEG): container finished" podID="52c549f0-ea5d-4708-85a7-0d0a6b4ece4e" containerID="90b4d26e4d3364abc0a61b52c21177fee27abafd7f691f7e37a082efe8b96309" exitCode=0 Oct 02 11:45:03 crc kubenswrapper[4771]: I1002 11:45:03.100894 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" event={"ID":"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e","Type":"ContainerDied","Data":"90b4d26e4d3364abc0a61b52c21177fee27abafd7f691f7e37a082efe8b96309"} Oct 02 11:45:04 crc kubenswrapper[4771]: I1002 11:45:04.360433 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pc9ng/crc-debug-d6lf7"] Oct 02 11:45:04 crc kubenswrapper[4771]: I1002 11:45:04.373840 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pc9ng/crc-debug-d6lf7"] Oct 02 11:45:04 crc kubenswrapper[4771]: I1002 11:45:04.581151 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" Oct 02 11:45:04 crc kubenswrapper[4771]: I1002 11:45:04.613579 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lg8z6\" (UniqueName: \"kubernetes.io/projected/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-kube-api-access-lg8z6\") pod \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\" (UID: \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\") " Oct 02 11:45:04 crc kubenswrapper[4771]: I1002 11:45:04.613710 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-config-volume\") pod \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\" (UID: \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\") " Oct 02 11:45:04 crc kubenswrapper[4771]: I1002 11:45:04.613851 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-secret-volume\") pod \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\" (UID: \"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e\") " Oct 02 11:45:04 crc kubenswrapper[4771]: I1002 11:45:04.614609 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-config-volume" (OuterVolumeSpecName: "config-volume") pod "52c549f0-ea5d-4708-85a7-0d0a6b4ece4e" (UID: "52c549f0-ea5d-4708-85a7-0d0a6b4ece4e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:45:04 crc kubenswrapper[4771]: I1002 11:45:04.620965 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "52c549f0-ea5d-4708-85a7-0d0a6b4ece4e" (UID: "52c549f0-ea5d-4708-85a7-0d0a6b4ece4e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:45:04 crc kubenswrapper[4771]: I1002 11:45:04.621808 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-kube-api-access-lg8z6" (OuterVolumeSpecName: "kube-api-access-lg8z6") pod "52c549f0-ea5d-4708-85a7-0d0a6b4ece4e" (UID: "52c549f0-ea5d-4708-85a7-0d0a6b4ece4e"). InnerVolumeSpecName "kube-api-access-lg8z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:45:04 crc kubenswrapper[4771]: I1002 11:45:04.717625 4771 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:04 crc kubenswrapper[4771]: I1002 11:45:04.717659 4771 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:04 crc kubenswrapper[4771]: I1002 11:45:04.717670 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lg8z6\" (UniqueName: \"kubernetes.io/projected/52c549f0-ea5d-4708-85a7-0d0a6b4ece4e-kube-api-access-lg8z6\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.127868 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" event={"ID":"52c549f0-ea5d-4708-85a7-0d0a6b4ece4e","Type":"ContainerDied","Data":"ab80705bb90f0935dce90738afa682ea196e80e2dc99e17624fffebc05a82f53"} Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.128214 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab80705bb90f0935dce90738afa682ea196e80e2dc99e17624fffebc05a82f53" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.127928 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-dwdjd" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.191862 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9"] Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.201605 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323380-8zld9"] Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.544978 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pc9ng/crc-debug-pfkz7"] Oct 02 11:45:05 crc kubenswrapper[4771]: E1002 11:45:05.545606 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52c549f0-ea5d-4708-85a7-0d0a6b4ece4e" containerName="collect-profiles" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.545622 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="52c549f0-ea5d-4708-85a7-0d0a6b4ece4e" containerName="collect-profiles" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.545994 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="52c549f0-ea5d-4708-85a7-0d0a6b4ece4e" containerName="collect-profiles" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.546860 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/crc-debug-pfkz7" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.639787 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f-host\") pod \"crc-debug-pfkz7\" (UID: \"b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f\") " pod="openshift-must-gather-pc9ng/crc-debug-pfkz7" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.640284 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvtpz\" (UniqueName: \"kubernetes.io/projected/b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f-kube-api-access-lvtpz\") pod \"crc-debug-pfkz7\" (UID: \"b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f\") " pod="openshift-must-gather-pc9ng/crc-debug-pfkz7" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.697517 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61229121-9b45-47d2-858c-5cc345408fc6" path="/var/lib/kubelet/pods/61229121-9b45-47d2-858c-5cc345408fc6/volumes" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.700440 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a89c2bce-3833-4930-9ac9-b82352e42f0f" path="/var/lib/kubelet/pods/a89c2bce-3833-4930-9ac9-b82352e42f0f/volumes" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.743043 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f-host\") pod \"crc-debug-pfkz7\" (UID: \"b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f\") " pod="openshift-must-gather-pc9ng/crc-debug-pfkz7" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.743354 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvtpz\" (UniqueName: \"kubernetes.io/projected/b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f-kube-api-access-lvtpz\") pod \"crc-debug-pfkz7\" (UID: \"b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f\") " pod="openshift-must-gather-pc9ng/crc-debug-pfkz7" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.744465 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f-host\") pod \"crc-debug-pfkz7\" (UID: \"b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f\") " pod="openshift-must-gather-pc9ng/crc-debug-pfkz7" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.770394 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvtpz\" (UniqueName: \"kubernetes.io/projected/b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f-kube-api-access-lvtpz\") pod \"crc-debug-pfkz7\" (UID: \"b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f\") " pod="openshift-must-gather-pc9ng/crc-debug-pfkz7" Oct 02 11:45:05 crc kubenswrapper[4771]: I1002 11:45:05.867723 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/crc-debug-pfkz7" Oct 02 11:45:05 crc kubenswrapper[4771]: W1002 11:45:05.918641 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6d23db3_4aa1_42c5_8b6d_1cc7c0ca882f.slice/crio-f1c105a2855f6b7d63e91a1372df71014253556afe48eb496a10566436d1214e WatchSource:0}: Error finding container f1c105a2855f6b7d63e91a1372df71014253556afe48eb496a10566436d1214e: Status 404 returned error can't find the container with id f1c105a2855f6b7d63e91a1372df71014253556afe48eb496a10566436d1214e Oct 02 11:45:06 crc kubenswrapper[4771]: I1002 11:45:06.140592 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pc9ng/crc-debug-pfkz7" event={"ID":"b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f","Type":"ContainerStarted","Data":"f1c105a2855f6b7d63e91a1372df71014253556afe48eb496a10566436d1214e"} Oct 02 11:45:07 crc kubenswrapper[4771]: I1002 11:45:07.154094 4771 generic.go:334] "Generic (PLEG): container finished" podID="b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f" containerID="be61aa44ce122b6ca29dedf8b60ce5cc7396aafc84a82cf7fc395069699805cb" exitCode=0 Oct 02 11:45:07 crc kubenswrapper[4771]: I1002 11:45:07.154187 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pc9ng/crc-debug-pfkz7" event={"ID":"b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f","Type":"ContainerDied","Data":"be61aa44ce122b6ca29dedf8b60ce5cc7396aafc84a82cf7fc395069699805cb"} Oct 02 11:45:07 crc kubenswrapper[4771]: I1002 11:45:07.198698 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pc9ng/crc-debug-pfkz7"] Oct 02 11:45:07 crc kubenswrapper[4771]: I1002 11:45:07.212598 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pc9ng/crc-debug-pfkz7"] Oct 02 11:45:08 crc kubenswrapper[4771]: I1002 11:45:08.041457 4771 scope.go:117] "RemoveContainer" containerID="c8bf9cf96c4f0287badd4798f71f35912aa0b9463fcfe5ce7f49b1f3f569a533" Oct 02 11:45:08 crc kubenswrapper[4771]: I1002 11:45:08.074622 4771 scope.go:117] "RemoveContainer" containerID="64ec96cfbdcfd553c58d753eb88e579b36aa4ef2f2019aea2268ffede9ca26a8" Oct 02 11:45:08 crc kubenswrapper[4771]: I1002 11:45:08.200536 4771 scope.go:117] "RemoveContainer" containerID="b1b02dba0c204955c89b53a6cd133b8183e67ac1d3f1a04f8b8ab854bcac7bff" Oct 02 11:45:08 crc kubenswrapper[4771]: I1002 11:45:08.316248 4771 scope.go:117] "RemoveContainer" containerID="20be41fd7d9d03d06bdc7d7626740fc68289635538c3e8752dff11402627ac72" Oct 02 11:45:08 crc kubenswrapper[4771]: I1002 11:45:08.371876 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/crc-debug-pfkz7" Oct 02 11:45:08 crc kubenswrapper[4771]: I1002 11:45:08.532564 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f-host\") pod \"b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f\" (UID: \"b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f\") " Oct 02 11:45:08 crc kubenswrapper[4771]: I1002 11:45:08.532770 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f-host" (OuterVolumeSpecName: "host") pod "b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f" (UID: "b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:45:08 crc kubenswrapper[4771]: I1002 11:45:08.533142 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvtpz\" (UniqueName: \"kubernetes.io/projected/b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f-kube-api-access-lvtpz\") pod \"b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f\" (UID: \"b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f\") " Oct 02 11:45:08 crc kubenswrapper[4771]: I1002 11:45:08.534516 4771 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f-host\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:08 crc kubenswrapper[4771]: I1002 11:45:08.539869 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f-kube-api-access-lvtpz" (OuterVolumeSpecName: "kube-api-access-lvtpz") pod "b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f" (UID: "b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f"). InnerVolumeSpecName "kube-api-access-lvtpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:45:08 crc kubenswrapper[4771]: I1002 11:45:08.636998 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvtpz\" (UniqueName: \"kubernetes.io/projected/b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f-kube-api-access-lvtpz\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:09 crc kubenswrapper[4771]: I1002 11:45:09.233310 4771 scope.go:117] "RemoveContainer" containerID="be61aa44ce122b6ca29dedf8b60ce5cc7396aafc84a82cf7fc395069699805cb" Oct 02 11:45:09 crc kubenswrapper[4771]: I1002 11:45:09.233336 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/crc-debug-pfkz7" Oct 02 11:45:09 crc kubenswrapper[4771]: I1002 11:45:09.481999 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq_08da521d-cddc-4af0-922b-673e392a68f3/util/0.log" Oct 02 11:45:09 crc kubenswrapper[4771]: I1002 11:45:09.702607 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f" path="/var/lib/kubelet/pods/b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f/volumes" Oct 02 11:45:09 crc kubenswrapper[4771]: I1002 11:45:09.722782 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq_08da521d-cddc-4af0-922b-673e392a68f3/util/0.log" Oct 02 11:45:09 crc kubenswrapper[4771]: I1002 11:45:09.790973 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq_08da521d-cddc-4af0-922b-673e392a68f3/pull/0.log" Oct 02 11:45:09 crc kubenswrapper[4771]: I1002 11:45:09.845884 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq_08da521d-cddc-4af0-922b-673e392a68f3/pull/0.log" Oct 02 11:45:10 crc kubenswrapper[4771]: I1002 11:45:10.043723 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq_08da521d-cddc-4af0-922b-673e392a68f3/util/0.log" Oct 02 11:45:10 crc kubenswrapper[4771]: I1002 11:45:10.069579 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq_08da521d-cddc-4af0-922b-673e392a68f3/extract/0.log" Oct 02 11:45:10 crc kubenswrapper[4771]: I1002 11:45:10.111660 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_838db991d96efe1746e91ce73faec15ca78c0f21648ab2feab8e4a730bxs8lq_08da521d-cddc-4af0-922b-673e392a68f3/pull/0.log" Oct 02 11:45:10 crc kubenswrapper[4771]: I1002 11:45:10.376205 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-dmwmw_5fcc6b1c-2163-454b-9172-16852b9f8966/kube-rbac-proxy/0.log" Oct 02 11:45:10 crc kubenswrapper[4771]: I1002 11:45:10.387989 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-smc8h_4da5dda4-f0da-453d-ba80-a2fc95035688/manager/0.log" Oct 02 11:45:10 crc kubenswrapper[4771]: I1002 11:45:10.424341 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-smc8h_4da5dda4-f0da-453d-ba80-a2fc95035688/kube-rbac-proxy/0.log" Oct 02 11:45:10 crc kubenswrapper[4771]: I1002 11:45:10.653623 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-dmwmw_5fcc6b1c-2163-454b-9172-16852b9f8966/manager/0.log" Oct 02 11:45:10 crc kubenswrapper[4771]: I1002 11:45:10.683611 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-5xvqw_f0638dc6-d014-4a78-bcfa-eb4e5cea5caf/kube-rbac-proxy/0.log" Oct 02 11:45:10 crc kubenswrapper[4771]: I1002 11:45:10.827763 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-5xvqw_f0638dc6-d014-4a78-bcfa-eb4e5cea5caf/manager/0.log" Oct 02 11:45:10 crc kubenswrapper[4771]: I1002 11:45:10.987748 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-pnqnd_6dc77080-a5ab-4db4-b150-2573589da332/kube-rbac-proxy/0.log" Oct 02 11:45:11 crc kubenswrapper[4771]: I1002 11:45:11.082448 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-pnqnd_6dc77080-a5ab-4db4-b150-2573589da332/manager/0.log" Oct 02 11:45:11 crc kubenswrapper[4771]: I1002 11:45:11.154723 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-hh756_9804f0f6-0273-4b09-830f-f082cf718803/kube-rbac-proxy/0.log" Oct 02 11:45:11 crc kubenswrapper[4771]: I1002 11:45:11.300177 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-hh756_9804f0f6-0273-4b09-830f-f082cf718803/manager/0.log" Oct 02 11:45:11 crc kubenswrapper[4771]: I1002 11:45:11.407827 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-lp8xj_8e8d717b-41cc-4dd7-b9e3-4d06ac36cc54/kube-rbac-proxy/0.log" Oct 02 11:45:11 crc kubenswrapper[4771]: I1002 11:45:11.421369 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-lp8xj_8e8d717b-41cc-4dd7-b9e3-4d06ac36cc54/manager/0.log" Oct 02 11:45:11 crc kubenswrapper[4771]: I1002 11:45:11.602079 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-99mgr_e0df3275-79e0-49b8-93d5-585e2667a5c4/kube-rbac-proxy/0.log" Oct 02 11:45:11 crc kubenswrapper[4771]: I1002 11:45:11.889687 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-99mgr_e0df3275-79e0-49b8-93d5-585e2667a5c4/manager/0.log" Oct 02 11:45:11 crc kubenswrapper[4771]: I1002 11:45:11.893461 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-4nr9l_7e625416-21a2-45e6-aec5-cf04c408e65e/manager/0.log" Oct 02 11:45:11 crc kubenswrapper[4771]: I1002 11:45:11.915592 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-4nr9l_7e625416-21a2-45e6-aec5-cf04c408e65e/kube-rbac-proxy/0.log" Oct 02 11:45:12 crc kubenswrapper[4771]: I1002 11:45:12.109424 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-jg6rv_1f236f0a-744a-4f98-90e9-7c78ead31ddc/kube-rbac-proxy/0.log" Oct 02 11:45:12 crc kubenswrapper[4771]: I1002 11:45:12.196998 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-jg6rv_1f236f0a-744a-4f98-90e9-7c78ead31ddc/manager/0.log" Oct 02 11:45:12 crc kubenswrapper[4771]: I1002 11:45:12.310149 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-4dcd9_04a0b3f5-6e78-477a-8bd6-9250ba41eabf/kube-rbac-proxy/0.log" Oct 02 11:45:12 crc kubenswrapper[4771]: I1002 11:45:12.358884 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-4dcd9_04a0b3f5-6e78-477a-8bd6-9250ba41eabf/manager/0.log" Oct 02 11:45:12 crc kubenswrapper[4771]: I1002 11:45:12.407254 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-tgx7p_7c2eb266-7f51-45b9-8095-89d219bb59f0/kube-rbac-proxy/0.log" Oct 02 11:45:12 crc kubenswrapper[4771]: I1002 11:45:12.632772 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-tgx7p_7c2eb266-7f51-45b9-8095-89d219bb59f0/manager/0.log" Oct 02 11:45:12 crc kubenswrapper[4771]: I1002 11:45:12.671919 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-zwjcd_4a399512-1d85-4f61-b8d4-800a57c33f4d/kube-rbac-proxy/0.log" Oct 02 11:45:12 crc kubenswrapper[4771]: I1002 11:45:12.861192 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-zwjcd_4a399512-1d85-4f61-b8d4-800a57c33f4d/manager/0.log" Oct 02 11:45:12 crc kubenswrapper[4771]: I1002 11:45:12.889519 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-mrfv5_9b273823-7cb9-4c85-b5cc-54a52a5d04b8/kube-rbac-proxy/0.log" Oct 02 11:45:13 crc kubenswrapper[4771]: I1002 11:45:13.053246 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-mrfv5_9b273823-7cb9-4c85-b5cc-54a52a5d04b8/manager/0.log" Oct 02 11:45:13 crc kubenswrapper[4771]: I1002 11:45:13.097471 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-4f4vk_4442fde9-3b77-4013-a6c2-87cfeca2b299/kube-rbac-proxy/0.log" Oct 02 11:45:13 crc kubenswrapper[4771]: I1002 11:45:13.098828 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-4f4vk_4442fde9-3b77-4013-a6c2-87cfeca2b299/manager/0.log" Oct 02 11:45:13 crc kubenswrapper[4771]: I1002 11:45:13.302557 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-lpb8d_ffc86aad-8b08-40c7-85f9-fb498bbf56b2/kube-rbac-proxy/0.log" Oct 02 11:45:13 crc kubenswrapper[4771]: I1002 11:45:13.305253 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-lpb8d_ffc86aad-8b08-40c7-85f9-fb498bbf56b2/manager/0.log" Oct 02 11:45:13 crc kubenswrapper[4771]: I1002 11:45:13.541599 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6567974949-gkcw5_6925b4e5-a255-4f79-b833-8862a67dc0ea/kube-rbac-proxy/0.log" Oct 02 11:45:13 crc kubenswrapper[4771]: I1002 11:45:13.620429 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-744d8c869b-vknn7_f01b0e6f-9cfa-45ac-878d-4af22bea031c/kube-rbac-proxy/0.log" Oct 02 11:45:13 crc kubenswrapper[4771]: I1002 11:45:13.881788 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-744d8c869b-vknn7_f01b0e6f-9cfa-45ac-878d-4af22bea031c/operator/0.log" Oct 02 11:45:13 crc kubenswrapper[4771]: I1002 11:45:13.888224 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-cpqp2_46c1bd11-aba5-4bc3-ae7e-fa24dd6687f6/registry-server/0.log" Oct 02 11:45:14 crc kubenswrapper[4771]: I1002 11:45:14.047963 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-gs7d5_ac8e19b1-242e-4785-96ee-4c72ee975cce/kube-rbac-proxy/0.log" Oct 02 11:45:14 crc kubenswrapper[4771]: I1002 11:45:14.208241 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-gs7d5_ac8e19b1-242e-4785-96ee-4c72ee975cce/manager/0.log" Oct 02 11:45:14 crc kubenswrapper[4771]: I1002 11:45:14.262986 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-l76nz_57aab528-153f-4c43-95d6-0c076d2071df/kube-rbac-proxy/0.log" Oct 02 11:45:14 crc kubenswrapper[4771]: I1002 11:45:14.287325 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-l76nz_57aab528-153f-4c43-95d6-0c076d2071df/manager/0.log" Oct 02 11:45:14 crc kubenswrapper[4771]: I1002 11:45:14.481260 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-7qk9l_ca4a791e-d59a-47c3-8f39-8d552c463c64/operator/0.log" Oct 02 11:45:14 crc kubenswrapper[4771]: I1002 11:45:14.527296 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-hxpcp_7f64d01d-a2f3-4e96-9f81-13cafb80225e/kube-rbac-proxy/0.log" Oct 02 11:45:14 crc kubenswrapper[4771]: I1002 11:45:14.685194 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-hxpcp_7f64d01d-a2f3-4e96-9f81-13cafb80225e/manager/0.log" Oct 02 11:45:14 crc kubenswrapper[4771]: I1002 11:45:14.721961 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d4856d54-q8sbj_a8a3258c-981f-4f61-bb08-42b5065a9067/kube-rbac-proxy/0.log" Oct 02 11:45:14 crc kubenswrapper[4771]: I1002 11:45:14.931818 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-59m2p_5c5a6559-fbef-41e7-bd03-71e40ef7ff7b/manager/0.log" Oct 02 11:45:14 crc kubenswrapper[4771]: I1002 11:45:14.950132 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-59m2p_5c5a6559-fbef-41e7-bd03-71e40ef7ff7b/kube-rbac-proxy/0.log" Oct 02 11:45:15 crc kubenswrapper[4771]: I1002 11:45:15.207619 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-wmrf6_f43ba614-8bea-47db-be7d-10c7a4e9ddbb/kube-rbac-proxy/0.log" Oct 02 11:45:15 crc kubenswrapper[4771]: I1002 11:45:15.233548 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-wmrf6_f43ba614-8bea-47db-be7d-10c7a4e9ddbb/manager/0.log" Oct 02 11:45:15 crc kubenswrapper[4771]: I1002 11:45:15.382557 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d4856d54-q8sbj_a8a3258c-981f-4f61-bb08-42b5065a9067/manager/0.log" Oct 02 11:45:15 crc kubenswrapper[4771]: I1002 11:45:15.428236 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6567974949-gkcw5_6925b4e5-a255-4f79-b833-8862a67dc0ea/manager/0.log" Oct 02 11:45:31 crc kubenswrapper[4771]: I1002 11:45:31.739021 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-xbvnm_61b3178b-0376-4db5-8369-af20d1a30056/control-plane-machine-set-operator/0.log" Oct 02 11:45:31 crc kubenswrapper[4771]: I1002 11:45:31.928326 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-68c2v_1cfc1793-db10-47a8-a152-0c86cc82be4c/kube-rbac-proxy/0.log" Oct 02 11:45:31 crc kubenswrapper[4771]: I1002 11:45:31.947956 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-68c2v_1cfc1793-db10-47a8-a152-0c86cc82be4c/machine-api-operator/0.log" Oct 02 11:45:42 crc kubenswrapper[4771]: I1002 11:45:42.145873 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:45:42 crc kubenswrapper[4771]: I1002 11:45:42.146365 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:45:44 crc kubenswrapper[4771]: I1002 11:45:44.346637 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-n9xk9_1eb44788-9a5a-44d1-9b40-909214b5b5f0/cert-manager-controller/0.log" Oct 02 11:45:44 crc kubenswrapper[4771]: I1002 11:45:44.482219 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-z9xqq_0d1d36a5-425d-4694-986f-87a45739053a/cert-manager-cainjector/0.log" Oct 02 11:45:44 crc kubenswrapper[4771]: I1002 11:45:44.514496 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-rmb84_38e28d25-0d2d-44b8-beb7-e5636e2dfdcd/cert-manager-webhook/0.log" Oct 02 11:45:58 crc kubenswrapper[4771]: I1002 11:45:58.442627 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-nlhsl_3de3b126-5868-413f-840f-f1502a0182e8/nmstate-console-plugin/0.log" Oct 02 11:45:58 crc kubenswrapper[4771]: I1002 11:45:58.638572 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-jxnvz_4b03aaef-425d-4ad0-85a2-04b3416b30e4/nmstate-handler/0.log" Oct 02 11:45:58 crc kubenswrapper[4771]: I1002 11:45:58.714030 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-2cnlz_95e5893b-430f-450e-8e67-44e0582a1583/kube-rbac-proxy/0.log" Oct 02 11:45:58 crc kubenswrapper[4771]: I1002 11:45:58.765392 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-2cnlz_95e5893b-430f-450e-8e67-44e0582a1583/nmstate-metrics/0.log" Oct 02 11:45:58 crc kubenswrapper[4771]: I1002 11:45:58.898795 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-lwrtk_0534552f-6859-4588-89ef-b2e5a036102b/nmstate-operator/0.log" Oct 02 11:45:59 crc kubenswrapper[4771]: I1002 11:45:59.036841 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-9gwrz_291a7779-68be-4198-8171-eab959730ce4/nmstate-webhook/0.log" Oct 02 11:46:12 crc kubenswrapper[4771]: I1002 11:46:12.145740 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:46:12 crc kubenswrapper[4771]: I1002 11:46:12.146143 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:46:12 crc kubenswrapper[4771]: I1002 11:46:12.745360 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6584497fb7-fcntv_063c19c9-6772-482a-bcbe-b1c620aac916/kube-rbac-proxy/0.log" Oct 02 11:46:12 crc kubenswrapper[4771]: I1002 11:46:12.764331 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6584497fb7-fcntv_063c19c9-6772-482a-bcbe-b1c620aac916/manager/0.log" Oct 02 11:46:27 crc kubenswrapper[4771]: I1002 11:46:27.976297 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-8958c8b87-r9qsk_7d4d0fad-fc61-4f27-b989-be4651af1895/cluster-logging-operator/0.log" Oct 02 11:46:28 crc kubenswrapper[4771]: I1002 11:46:28.168661 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-llprk_0d5fb871-be70-4492-b6c9-6f7b6684b890/collector/0.log" Oct 02 11:46:28 crc kubenswrapper[4771]: I1002 11:46:28.281580 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_41d40d08-10d1-4abc-90e4-f894d3ccfa4b/loki-compactor/0.log" Oct 02 11:46:28 crc kubenswrapper[4771]: I1002 11:46:28.410591 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-6f5f7fff97-47hsj_e9014359-f9dc-4e80-83c1-127149a3a60e/loki-distributor/0.log" Oct 02 11:46:28 crc kubenswrapper[4771]: I1002 11:46:28.530277 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-6dc97c5fb-2nswh_21daf834-8ac7-4137-933e-b7d3db27e225/gateway/0.log" Oct 02 11:46:28 crc kubenswrapper[4771]: I1002 11:46:28.566342 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-6dc97c5fb-2nswh_21daf834-8ac7-4137-933e-b7d3db27e225/opa/0.log" Oct 02 11:46:28 crc kubenswrapper[4771]: I1002 11:46:28.738767 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-6dc97c5fb-vvdff_f9c8afd6-efbc-4787-89ed-100438b03ecd/opa/0.log" Oct 02 11:46:28 crc kubenswrapper[4771]: I1002 11:46:28.785106 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-6dc97c5fb-vvdff_f9c8afd6-efbc-4787-89ed-100438b03ecd/gateway/0.log" Oct 02 11:46:28 crc kubenswrapper[4771]: I1002 11:46:28.880463 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_c00667ae-ca81-4c00-9ab2-71a1f307978d/loki-index-gateway/0.log" Oct 02 11:46:29 crc kubenswrapper[4771]: I1002 11:46:29.085003 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_34f9375e-2699-41ae-a604-81599cc16ea3/loki-ingester/0.log" Oct 02 11:46:29 crc kubenswrapper[4771]: I1002 11:46:29.128755 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5d954896cf-74tjf_aca1f6a0-c677-4129-84ec-0d33b51193d2/loki-querier/0.log" Oct 02 11:46:29 crc kubenswrapper[4771]: I1002 11:46:29.260218 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-6fbbbc8b7d-twxfl_cdb0a9ad-c8af-4298-9726-fa21864ee2a6/loki-query-frontend/0.log" Oct 02 11:46:42 crc kubenswrapper[4771]: I1002 11:46:42.145913 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:46:42 crc kubenswrapper[4771]: I1002 11:46:42.146486 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:46:42 crc kubenswrapper[4771]: I1002 11:46:42.146538 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 11:46:42 crc kubenswrapper[4771]: I1002 11:46:42.147616 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:46:42 crc kubenswrapper[4771]: I1002 11:46:42.147677 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" gracePeriod=600 Oct 02 11:46:42 crc kubenswrapper[4771]: E1002 11:46:42.286765 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:46:42 crc kubenswrapper[4771]: I1002 11:46:42.972535 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-hsbdv_37d57598-c98a-45ff-8de0-c51a5e39adeb/kube-rbac-proxy/0.log" Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.205356 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/cp-frr-files/0.log" Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.262215 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" exitCode=0 Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.262268 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5"} Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.262308 4771 scope.go:117] "RemoveContainer" containerID="3bb74c9f46be75efdbfc800b6d2f38f7da79e737ae4bf6020f5528c548db4dc6" Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.263990 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:46:43 crc kubenswrapper[4771]: E1002 11:46:43.265028 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.269156 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-hsbdv_37d57598-c98a-45ff-8de0-c51a5e39adeb/controller/0.log" Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.455904 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/cp-frr-files/0.log" Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.527444 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/cp-metrics/0.log" Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.532180 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/cp-reloader/0.log" Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.596420 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/cp-reloader/0.log" Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.702555 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/cp-frr-files/0.log" Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.742867 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/cp-reloader/0.log" Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.743983 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/cp-metrics/0.log" Oct 02 11:46:43 crc kubenswrapper[4771]: I1002 11:46:43.831740 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/cp-metrics/0.log" Oct 02 11:46:44 crc kubenswrapper[4771]: I1002 11:46:44.034075 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/cp-metrics/0.log" Oct 02 11:46:44 crc kubenswrapper[4771]: I1002 11:46:44.055163 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/cp-reloader/0.log" Oct 02 11:46:44 crc kubenswrapper[4771]: I1002 11:46:44.104685 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/controller/0.log" Oct 02 11:46:44 crc kubenswrapper[4771]: I1002 11:46:44.109207 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/cp-frr-files/0.log" Oct 02 11:46:44 crc kubenswrapper[4771]: I1002 11:46:44.311480 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/frr-metrics/0.log" Oct 02 11:46:44 crc kubenswrapper[4771]: I1002 11:46:44.345205 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/kube-rbac-proxy/0.log" Oct 02 11:46:44 crc kubenswrapper[4771]: I1002 11:46:44.394153 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/kube-rbac-proxy-frr/0.log" Oct 02 11:46:44 crc kubenswrapper[4771]: I1002 11:46:44.532226 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/reloader/0.log" Oct 02 11:46:44 crc kubenswrapper[4771]: I1002 11:46:44.703705 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-qvlg4_3c832819-5b80-4548-8941-6e411bd1ba2f/frr-k8s-webhook-server/0.log" Oct 02 11:46:44 crc kubenswrapper[4771]: I1002 11:46:44.854620 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7bc8798995-4p5ck_6ff4fcd7-2222-4405-911d-6d337e1af8a2/manager/0.log" Oct 02 11:46:44 crc kubenswrapper[4771]: I1002 11:46:44.997879 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-8547c6bb6c-bjb6j_8968c764-9017-43a2-bc9e-c2d3fe04a7fc/webhook-server/0.log" Oct 02 11:46:45 crc kubenswrapper[4771]: I1002 11:46:45.246614 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nrj2m_381993c1-33fd-485c-bb38-3f18cfa91abf/kube-rbac-proxy/0.log" Oct 02 11:46:46 crc kubenswrapper[4771]: I1002 11:46:46.384211 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nrj2m_381993c1-33fd-485c-bb38-3f18cfa91abf/speaker/0.log" Oct 02 11:46:46 crc kubenswrapper[4771]: I1002 11:46:46.685308 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cf6h7_5638abcc-fdb8-4ed6-a630-8562d79668a5/frr/0.log" Oct 02 11:46:57 crc kubenswrapper[4771]: I1002 11:46:57.681992 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:46:57 crc kubenswrapper[4771]: E1002 11:46:57.682830 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:46:58 crc kubenswrapper[4771]: I1002 11:46:58.491529 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf_64cf8242-2c71-4831-b744-679193ae1ec6/util/0.log" Oct 02 11:46:58 crc kubenswrapper[4771]: I1002 11:46:58.741618 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf_64cf8242-2c71-4831-b744-679193ae1ec6/pull/0.log" Oct 02 11:46:58 crc kubenswrapper[4771]: I1002 11:46:58.751826 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf_64cf8242-2c71-4831-b744-679193ae1ec6/util/0.log" Oct 02 11:46:58 crc kubenswrapper[4771]: I1002 11:46:58.789488 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf_64cf8242-2c71-4831-b744-679193ae1ec6/pull/0.log" Oct 02 11:46:58 crc kubenswrapper[4771]: I1002 11:46:58.965714 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf_64cf8242-2c71-4831-b744-679193ae1ec6/extract/0.log" Oct 02 11:46:59 crc kubenswrapper[4771]: I1002 11:46:59.007886 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf_64cf8242-2c71-4831-b744-679193ae1ec6/pull/0.log" Oct 02 11:46:59 crc kubenswrapper[4771]: I1002 11:46:59.008676 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0999b653a84702c2c11c13a5389e053aa7f0501a9a88eec9014235a37dt7vtf_64cf8242-2c71-4831-b744-679193ae1ec6/util/0.log" Oct 02 11:46:59 crc kubenswrapper[4771]: I1002 11:46:59.145125 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5_0e20c875-b90e-424e-9fe3-e873406a86db/util/0.log" Oct 02 11:46:59 crc kubenswrapper[4771]: I1002 11:46:59.327044 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5_0e20c875-b90e-424e-9fe3-e873406a86db/util/0.log" Oct 02 11:46:59 crc kubenswrapper[4771]: I1002 11:46:59.338175 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5_0e20c875-b90e-424e-9fe3-e873406a86db/pull/0.log" Oct 02 11:46:59 crc kubenswrapper[4771]: I1002 11:46:59.339361 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5_0e20c875-b90e-424e-9fe3-e873406a86db/pull/0.log" Oct 02 11:46:59 crc kubenswrapper[4771]: I1002 11:46:59.535918 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5_0e20c875-b90e-424e-9fe3-e873406a86db/pull/0.log" Oct 02 11:46:59 crc kubenswrapper[4771]: I1002 11:46:59.576122 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5_0e20c875-b90e-424e-9fe3-e873406a86db/extract/0.log" Oct 02 11:46:59 crc kubenswrapper[4771]: I1002 11:46:59.579604 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mstp5_0e20c875-b90e-424e-9fe3-e873406a86db/util/0.log" Oct 02 11:46:59 crc kubenswrapper[4771]: I1002 11:46:59.704294 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g_238c1703-8f29-422d-8fcf-cb1498b81108/util/0.log" Oct 02 11:46:59 crc kubenswrapper[4771]: I1002 11:46:59.935276 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g_238c1703-8f29-422d-8fcf-cb1498b81108/util/0.log" Oct 02 11:46:59 crc kubenswrapper[4771]: I1002 11:46:59.942966 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g_238c1703-8f29-422d-8fcf-cb1498b81108/pull/0.log" Oct 02 11:47:00 crc kubenswrapper[4771]: I1002 11:47:00.016240 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g_238c1703-8f29-422d-8fcf-cb1498b81108/pull/0.log" Oct 02 11:47:00 crc kubenswrapper[4771]: I1002 11:47:00.155100 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g_238c1703-8f29-422d-8fcf-cb1498b81108/extract/0.log" Oct 02 11:47:00 crc kubenswrapper[4771]: I1002 11:47:00.187684 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g_238c1703-8f29-422d-8fcf-cb1498b81108/pull/0.log" Oct 02 11:47:00 crc kubenswrapper[4771]: I1002 11:47:00.193902 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dnkg8g_238c1703-8f29-422d-8fcf-cb1498b81108/util/0.log" Oct 02 11:47:00 crc kubenswrapper[4771]: I1002 11:47:00.337319 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x_d3c0ef21-3031-4032-89d8-001a64387d57/util/0.log" Oct 02 11:47:00 crc kubenswrapper[4771]: I1002 11:47:00.548305 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x_d3c0ef21-3031-4032-89d8-001a64387d57/util/0.log" Oct 02 11:47:00 crc kubenswrapper[4771]: I1002 11:47:00.552231 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x_d3c0ef21-3031-4032-89d8-001a64387d57/pull/0.log" Oct 02 11:47:00 crc kubenswrapper[4771]: I1002 11:47:00.552247 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x_d3c0ef21-3031-4032-89d8-001a64387d57/pull/0.log" Oct 02 11:47:00 crc kubenswrapper[4771]: I1002 11:47:00.726926 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x_d3c0ef21-3031-4032-89d8-001a64387d57/pull/0.log" Oct 02 11:47:00 crc kubenswrapper[4771]: I1002 11:47:00.746771 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x_d3c0ef21-3031-4032-89d8-001a64387d57/extract/0.log" Oct 02 11:47:00 crc kubenswrapper[4771]: I1002 11:47:00.751851 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_b750ce2fcb78a523ce3c4c91d54ad8430abb37e936593acebfbbbfa601g668x_d3c0ef21-3031-4032-89d8-001a64387d57/util/0.log" Oct 02 11:47:00 crc kubenswrapper[4771]: I1002 11:47:00.931626 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-md7x4_c89214fa-0d7c-4937-9cf3-6545a47420b9/extract-utilities/0.log" Oct 02 11:47:01 crc kubenswrapper[4771]: I1002 11:47:01.118086 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-md7x4_c89214fa-0d7c-4937-9cf3-6545a47420b9/extract-utilities/0.log" Oct 02 11:47:01 crc kubenswrapper[4771]: I1002 11:47:01.125193 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-md7x4_c89214fa-0d7c-4937-9cf3-6545a47420b9/extract-content/0.log" Oct 02 11:47:01 crc kubenswrapper[4771]: I1002 11:47:01.153401 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-md7x4_c89214fa-0d7c-4937-9cf3-6545a47420b9/extract-content/0.log" Oct 02 11:47:01 crc kubenswrapper[4771]: I1002 11:47:01.337927 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-md7x4_c89214fa-0d7c-4937-9cf3-6545a47420b9/extract-utilities/0.log" Oct 02 11:47:01 crc kubenswrapper[4771]: I1002 11:47:01.347670 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-md7x4_c89214fa-0d7c-4937-9cf3-6545a47420b9/extract-content/0.log" Oct 02 11:47:01 crc kubenswrapper[4771]: I1002 11:47:01.587425 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-grhhm_916b5469-2407-426e-ac06-157981e07018/extract-utilities/0.log" Oct 02 11:47:01 crc kubenswrapper[4771]: I1002 11:47:01.885011 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-grhhm_916b5469-2407-426e-ac06-157981e07018/extract-content/0.log" Oct 02 11:47:01 crc kubenswrapper[4771]: I1002 11:47:01.923938 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-grhhm_916b5469-2407-426e-ac06-157981e07018/extract-content/0.log" Oct 02 11:47:01 crc kubenswrapper[4771]: I1002 11:47:01.924008 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-grhhm_916b5469-2407-426e-ac06-157981e07018/extract-utilities/0.log" Oct 02 11:47:02 crc kubenswrapper[4771]: I1002 11:47:02.128896 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-grhhm_916b5469-2407-426e-ac06-157981e07018/extract-utilities/0.log" Oct 02 11:47:02 crc kubenswrapper[4771]: I1002 11:47:02.132861 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-grhhm_916b5469-2407-426e-ac06-157981e07018/extract-content/0.log" Oct 02 11:47:02 crc kubenswrapper[4771]: I1002 11:47:02.398258 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt_23531133-454e-4938-bcad-744b9ee79a26/util/0.log" Oct 02 11:47:02 crc kubenswrapper[4771]: I1002 11:47:02.739390 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt_23531133-454e-4938-bcad-744b9ee79a26/pull/0.log" Oct 02 11:47:02 crc kubenswrapper[4771]: I1002 11:47:02.745721 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt_23531133-454e-4938-bcad-744b9ee79a26/pull/0.log" Oct 02 11:47:02 crc kubenswrapper[4771]: I1002 11:47:02.776291 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt_23531133-454e-4938-bcad-744b9ee79a26/util/0.log" Oct 02 11:47:02 crc kubenswrapper[4771]: I1002 11:47:02.781641 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-md7x4_c89214fa-0d7c-4937-9cf3-6545a47420b9/registry-server/0.log" Oct 02 11:47:03 crc kubenswrapper[4771]: I1002 11:47:03.029161 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt_23531133-454e-4938-bcad-744b9ee79a26/extract/0.log" Oct 02 11:47:03 crc kubenswrapper[4771]: I1002 11:47:03.064721 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt_23531133-454e-4938-bcad-744b9ee79a26/pull/0.log" Oct 02 11:47:03 crc kubenswrapper[4771]: I1002 11:47:03.087471 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdcnwt_23531133-454e-4938-bcad-744b9ee79a26/util/0.log" Oct 02 11:47:03 crc kubenswrapper[4771]: I1002 11:47:03.337667 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-v7w6j_36a9a4c8-f9ed-4ed0-b46c-5bd1760e7a78/marketplace-operator/0.log" Oct 02 11:47:03 crc kubenswrapper[4771]: I1002 11:47:03.397693 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w8d6f_e228deeb-ac11-4ca5-8e08-5395e28bfb07/extract-utilities/0.log" Oct 02 11:47:03 crc kubenswrapper[4771]: I1002 11:47:03.574590 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w8d6f_e228deeb-ac11-4ca5-8e08-5395e28bfb07/extract-utilities/0.log" Oct 02 11:47:03 crc kubenswrapper[4771]: I1002 11:47:03.630018 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w8d6f_e228deeb-ac11-4ca5-8e08-5395e28bfb07/extract-content/0.log" Oct 02 11:47:03 crc kubenswrapper[4771]: I1002 11:47:03.630207 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w8d6f_e228deeb-ac11-4ca5-8e08-5395e28bfb07/extract-content/0.log" Oct 02 11:47:03 crc kubenswrapper[4771]: I1002 11:47:03.770353 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-grhhm_916b5469-2407-426e-ac06-157981e07018/registry-server/0.log" Oct 02 11:47:03 crc kubenswrapper[4771]: I1002 11:47:03.919400 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w8d6f_e228deeb-ac11-4ca5-8e08-5395e28bfb07/extract-content/0.log" Oct 02 11:47:03 crc kubenswrapper[4771]: I1002 11:47:03.975724 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w8d6f_e228deeb-ac11-4ca5-8e08-5395e28bfb07/extract-utilities/0.log" Oct 02 11:47:04 crc kubenswrapper[4771]: I1002 11:47:04.034698 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9sfxx_9be41903-7359-484f-8e0d-3a94e53ade9a/extract-utilities/0.log" Oct 02 11:47:04 crc kubenswrapper[4771]: I1002 11:47:04.185647 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-w8d6f_e228deeb-ac11-4ca5-8e08-5395e28bfb07/registry-server/0.log" Oct 02 11:47:04 crc kubenswrapper[4771]: I1002 11:47:04.236512 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9sfxx_9be41903-7359-484f-8e0d-3a94e53ade9a/extract-utilities/0.log" Oct 02 11:47:04 crc kubenswrapper[4771]: I1002 11:47:04.259265 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9sfxx_9be41903-7359-484f-8e0d-3a94e53ade9a/extract-content/0.log" Oct 02 11:47:04 crc kubenswrapper[4771]: I1002 11:47:04.261055 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9sfxx_9be41903-7359-484f-8e0d-3a94e53ade9a/extract-content/0.log" Oct 02 11:47:04 crc kubenswrapper[4771]: I1002 11:47:04.519439 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9sfxx_9be41903-7359-484f-8e0d-3a94e53ade9a/extract-utilities/0.log" Oct 02 11:47:04 crc kubenswrapper[4771]: I1002 11:47:04.529457 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9sfxx_9be41903-7359-484f-8e0d-3a94e53ade9a/extract-content/0.log" Oct 02 11:47:05 crc kubenswrapper[4771]: I1002 11:47:05.555003 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9sfxx_9be41903-7359-484f-8e0d-3a94e53ade9a/registry-server/0.log" Oct 02 11:47:08 crc kubenswrapper[4771]: I1002 11:47:08.682730 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:47:08 crc kubenswrapper[4771]: E1002 11:47:08.683684 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:47:17 crc kubenswrapper[4771]: I1002 11:47:17.052240 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-tbgps_9c0d8940-6aa3-461a-9b29-53277779bb47/prometheus-operator/0.log" Oct 02 11:47:17 crc kubenswrapper[4771]: I1002 11:47:17.222468 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-8dfcdd589-2knk2_6214e26e-87c0-4bfb-bb9f-889636a4a9fa/prometheus-operator-admission-webhook/0.log" Oct 02 11:47:17 crc kubenswrapper[4771]: I1002 11:47:17.265097 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-8dfcdd589-bzfwn_8fd0bf98-7976-4e81-8afa-76b37aff3944/prometheus-operator-admission-webhook/0.log" Oct 02 11:47:17 crc kubenswrapper[4771]: I1002 11:47:17.443306 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-wwgrl_e792e389-a48d-47fa-8c41-9767cb669865/operator/0.log" Oct 02 11:47:17 crc kubenswrapper[4771]: I1002 11:47:17.490558 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-6584dc9448-9vbgq_7636c0e3-830f-4f22-88ca-52a4e1358e08/observability-ui-dashboards/0.log" Oct 02 11:47:17 crc kubenswrapper[4771]: I1002 11:47:17.652285 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-2ccmn_fa786166-e5d7-43e7-8803-98135ffc2871/perses-operator/0.log" Oct 02 11:47:19 crc kubenswrapper[4771]: I1002 11:47:19.681957 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:47:19 crc kubenswrapper[4771]: E1002 11:47:19.682886 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:47:30 crc kubenswrapper[4771]: I1002 11:47:30.779741 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6584497fb7-fcntv_063c19c9-6772-482a-bcbe-b1c620aac916/manager/0.log" Oct 02 11:47:30 crc kubenswrapper[4771]: I1002 11:47:30.835826 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6584497fb7-fcntv_063c19c9-6772-482a-bcbe-b1c620aac916/kube-rbac-proxy/0.log" Oct 02 11:47:33 crc kubenswrapper[4771]: I1002 11:47:33.694683 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:47:33 crc kubenswrapper[4771]: E1002 11:47:33.700185 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:47:45 crc kubenswrapper[4771]: I1002 11:47:45.681859 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:47:45 crc kubenswrapper[4771]: E1002 11:47:45.682747 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:47:58 crc kubenswrapper[4771]: I1002 11:47:58.683823 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:47:58 crc kubenswrapper[4771]: E1002 11:47:58.684553 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.364601 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rmfcw"] Oct 02 11:48:00 crc kubenswrapper[4771]: E1002 11:48:00.365627 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f" containerName="container-00" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.365645 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f" containerName="container-00" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.366027 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d23db3-4aa1-42c5-8b6d-1cc7c0ca882f" containerName="container-00" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.368167 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.381172 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rmfcw"] Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.514432 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-catalog-content\") pod \"redhat-operators-rmfcw\" (UID: \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\") " pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.514837 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-utilities\") pod \"redhat-operators-rmfcw\" (UID: \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\") " pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.514920 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvkpp\" (UniqueName: \"kubernetes.io/projected/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-kube-api-access-cvkpp\") pod \"redhat-operators-rmfcw\" (UID: \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\") " pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.617451 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-catalog-content\") pod \"redhat-operators-rmfcw\" (UID: \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\") " pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.617915 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-utilities\") pod \"redhat-operators-rmfcw\" (UID: \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\") " pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.618025 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvkpp\" (UniqueName: \"kubernetes.io/projected/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-kube-api-access-cvkpp\") pod \"redhat-operators-rmfcw\" (UID: \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\") " pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.618560 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-catalog-content\") pod \"redhat-operators-rmfcw\" (UID: \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\") " pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.618756 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-utilities\") pod \"redhat-operators-rmfcw\" (UID: \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\") " pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.644842 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvkpp\" (UniqueName: \"kubernetes.io/projected/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-kube-api-access-cvkpp\") pod \"redhat-operators-rmfcw\" (UID: \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\") " pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:00 crc kubenswrapper[4771]: I1002 11:48:00.694786 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:01 crc kubenswrapper[4771]: I1002 11:48:01.341745 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rmfcw"] Oct 02 11:48:02 crc kubenswrapper[4771]: I1002 11:48:02.130558 4771 generic.go:334] "Generic (PLEG): container finished" podID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" containerID="b46e6b048159ef62e76eda6b3546ac01c6a55e2d5bf8c0541d9cad13aef7fd47" exitCode=0 Oct 02 11:48:02 crc kubenswrapper[4771]: I1002 11:48:02.131005 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmfcw" event={"ID":"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8","Type":"ContainerDied","Data":"b46e6b048159ef62e76eda6b3546ac01c6a55e2d5bf8c0541d9cad13aef7fd47"} Oct 02 11:48:02 crc kubenswrapper[4771]: I1002 11:48:02.131031 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmfcw" event={"ID":"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8","Type":"ContainerStarted","Data":"e9dcf0483d40e148beec765f53831f338b232337c5aff4c311bf85292583455f"} Oct 02 11:48:04 crc kubenswrapper[4771]: I1002 11:48:04.154956 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmfcw" event={"ID":"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8","Type":"ContainerStarted","Data":"0b8b5bf97829ac416c8a8e5d758e67c930c1df00c2f0841566d15dfe88b444fd"} Oct 02 11:48:08 crc kubenswrapper[4771]: I1002 11:48:08.201536 4771 generic.go:334] "Generic (PLEG): container finished" podID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" containerID="0b8b5bf97829ac416c8a8e5d758e67c930c1df00c2f0841566d15dfe88b444fd" exitCode=0 Oct 02 11:48:08 crc kubenswrapper[4771]: I1002 11:48:08.201609 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmfcw" event={"ID":"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8","Type":"ContainerDied","Data":"0b8b5bf97829ac416c8a8e5d758e67c930c1df00c2f0841566d15dfe88b444fd"} Oct 02 11:48:10 crc kubenswrapper[4771]: I1002 11:48:10.225650 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmfcw" event={"ID":"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8","Type":"ContainerStarted","Data":"4cfcf0d117ba7b009a2444b749ca94427f5c0428f32c7a86ea111adc9923dad0"} Oct 02 11:48:10 crc kubenswrapper[4771]: I1002 11:48:10.263936 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rmfcw" podStartSLOduration=2.52774755 podStartE2EDuration="10.263916479s" podCreationTimestamp="2025-10-02 11:48:00 +0000 UTC" firstStartedPulling="2025-10-02 11:48:02.133471778 +0000 UTC m=+7869.781156845" lastFinishedPulling="2025-10-02 11:48:09.869640717 +0000 UTC m=+7877.517325774" observedRunningTime="2025-10-02 11:48:10.251726864 +0000 UTC m=+7877.899411941" watchObservedRunningTime="2025-10-02 11:48:10.263916479 +0000 UTC m=+7877.911601546" Oct 02 11:48:10 crc kubenswrapper[4771]: I1002 11:48:10.695547 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:10 crc kubenswrapper[4771]: I1002 11:48:10.695632 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:11 crc kubenswrapper[4771]: I1002 11:48:11.683411 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:48:11 crc kubenswrapper[4771]: E1002 11:48:11.684250 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:48:11 crc kubenswrapper[4771]: I1002 11:48:11.750847 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rmfcw" podUID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" containerName="registry-server" probeResult="failure" output=< Oct 02 11:48:11 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:48:11 crc kubenswrapper[4771]: > Oct 02 11:48:21 crc kubenswrapper[4771]: I1002 11:48:21.756231 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rmfcw" podUID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" containerName="registry-server" probeResult="failure" output=< Oct 02 11:48:21 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:48:21 crc kubenswrapper[4771]: > Oct 02 11:48:24 crc kubenswrapper[4771]: I1002 11:48:24.683159 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:48:24 crc kubenswrapper[4771]: E1002 11:48:24.685563 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:48:30 crc kubenswrapper[4771]: I1002 11:48:30.745518 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:30 crc kubenswrapper[4771]: I1002 11:48:30.795544 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:30 crc kubenswrapper[4771]: I1002 11:48:30.983372 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rmfcw"] Oct 02 11:48:32 crc kubenswrapper[4771]: I1002 11:48:32.482622 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rmfcw" podUID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" containerName="registry-server" containerID="cri-o://4cfcf0d117ba7b009a2444b749ca94427f5c0428f32c7a86ea111adc9923dad0" gracePeriod=2 Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.167370 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.250543 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-utilities\") pod \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\" (UID: \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\") " Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.250732 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-catalog-content\") pod \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\" (UID: \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\") " Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.250784 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvkpp\" (UniqueName: \"kubernetes.io/projected/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-kube-api-access-cvkpp\") pod \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\" (UID: \"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8\") " Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.251352 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-utilities" (OuterVolumeSpecName: "utilities") pod "6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" (UID: "6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.260927 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-kube-api-access-cvkpp" (OuterVolumeSpecName: "kube-api-access-cvkpp") pod "6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" (UID: "6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8"). InnerVolumeSpecName "kube-api-access-cvkpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.329046 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" (UID: "6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.354551 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.354595 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.354606 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvkpp\" (UniqueName: \"kubernetes.io/projected/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8-kube-api-access-cvkpp\") on node \"crc\" DevicePath \"\"" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.495282 4771 generic.go:334] "Generic (PLEG): container finished" podID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" containerID="4cfcf0d117ba7b009a2444b749ca94427f5c0428f32c7a86ea111adc9923dad0" exitCode=0 Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.495350 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rmfcw" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.495359 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmfcw" event={"ID":"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8","Type":"ContainerDied","Data":"4cfcf0d117ba7b009a2444b749ca94427f5c0428f32c7a86ea111adc9923dad0"} Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.495778 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmfcw" event={"ID":"6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8","Type":"ContainerDied","Data":"e9dcf0483d40e148beec765f53831f338b232337c5aff4c311bf85292583455f"} Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.495802 4771 scope.go:117] "RemoveContainer" containerID="4cfcf0d117ba7b009a2444b749ca94427f5c0428f32c7a86ea111adc9923dad0" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.531726 4771 scope.go:117] "RemoveContainer" containerID="0b8b5bf97829ac416c8a8e5d758e67c930c1df00c2f0841566d15dfe88b444fd" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.535840 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rmfcw"] Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.552632 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rmfcw"] Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.559278 4771 scope.go:117] "RemoveContainer" containerID="b46e6b048159ef62e76eda6b3546ac01c6a55e2d5bf8c0541d9cad13aef7fd47" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.623918 4771 scope.go:117] "RemoveContainer" containerID="4cfcf0d117ba7b009a2444b749ca94427f5c0428f32c7a86ea111adc9923dad0" Oct 02 11:48:33 crc kubenswrapper[4771]: E1002 11:48:33.624558 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cfcf0d117ba7b009a2444b749ca94427f5c0428f32c7a86ea111adc9923dad0\": container with ID starting with 4cfcf0d117ba7b009a2444b749ca94427f5c0428f32c7a86ea111adc9923dad0 not found: ID does not exist" containerID="4cfcf0d117ba7b009a2444b749ca94427f5c0428f32c7a86ea111adc9923dad0" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.624598 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cfcf0d117ba7b009a2444b749ca94427f5c0428f32c7a86ea111adc9923dad0"} err="failed to get container status \"4cfcf0d117ba7b009a2444b749ca94427f5c0428f32c7a86ea111adc9923dad0\": rpc error: code = NotFound desc = could not find container \"4cfcf0d117ba7b009a2444b749ca94427f5c0428f32c7a86ea111adc9923dad0\": container with ID starting with 4cfcf0d117ba7b009a2444b749ca94427f5c0428f32c7a86ea111adc9923dad0 not found: ID does not exist" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.624637 4771 scope.go:117] "RemoveContainer" containerID="0b8b5bf97829ac416c8a8e5d758e67c930c1df00c2f0841566d15dfe88b444fd" Oct 02 11:48:33 crc kubenswrapper[4771]: E1002 11:48:33.625092 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b8b5bf97829ac416c8a8e5d758e67c930c1df00c2f0841566d15dfe88b444fd\": container with ID starting with 0b8b5bf97829ac416c8a8e5d758e67c930c1df00c2f0841566d15dfe88b444fd not found: ID does not exist" containerID="0b8b5bf97829ac416c8a8e5d758e67c930c1df00c2f0841566d15dfe88b444fd" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.625151 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8b5bf97829ac416c8a8e5d758e67c930c1df00c2f0841566d15dfe88b444fd"} err="failed to get container status \"0b8b5bf97829ac416c8a8e5d758e67c930c1df00c2f0841566d15dfe88b444fd\": rpc error: code = NotFound desc = could not find container \"0b8b5bf97829ac416c8a8e5d758e67c930c1df00c2f0841566d15dfe88b444fd\": container with ID starting with 0b8b5bf97829ac416c8a8e5d758e67c930c1df00c2f0841566d15dfe88b444fd not found: ID does not exist" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.625180 4771 scope.go:117] "RemoveContainer" containerID="b46e6b048159ef62e76eda6b3546ac01c6a55e2d5bf8c0541d9cad13aef7fd47" Oct 02 11:48:33 crc kubenswrapper[4771]: E1002 11:48:33.625504 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b46e6b048159ef62e76eda6b3546ac01c6a55e2d5bf8c0541d9cad13aef7fd47\": container with ID starting with b46e6b048159ef62e76eda6b3546ac01c6a55e2d5bf8c0541d9cad13aef7fd47 not found: ID does not exist" containerID="b46e6b048159ef62e76eda6b3546ac01c6a55e2d5bf8c0541d9cad13aef7fd47" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.625552 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b46e6b048159ef62e76eda6b3546ac01c6a55e2d5bf8c0541d9cad13aef7fd47"} err="failed to get container status \"b46e6b048159ef62e76eda6b3546ac01c6a55e2d5bf8c0541d9cad13aef7fd47\": rpc error: code = NotFound desc = could not find container \"b46e6b048159ef62e76eda6b3546ac01c6a55e2d5bf8c0541d9cad13aef7fd47\": container with ID starting with b46e6b048159ef62e76eda6b3546ac01c6a55e2d5bf8c0541d9cad13aef7fd47 not found: ID does not exist" Oct 02 11:48:33 crc kubenswrapper[4771]: I1002 11:48:33.695847 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" path="/var/lib/kubelet/pods/6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8/volumes" Oct 02 11:48:36 crc kubenswrapper[4771]: I1002 11:48:36.682259 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:48:36 crc kubenswrapper[4771]: E1002 11:48:36.683060 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:48:37 crc kubenswrapper[4771]: I1002 11:48:37.896094 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2hj8n"] Oct 02 11:48:37 crc kubenswrapper[4771]: E1002 11:48:37.896646 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" containerName="extract-utilities" Oct 02 11:48:37 crc kubenswrapper[4771]: I1002 11:48:37.896661 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" containerName="extract-utilities" Oct 02 11:48:37 crc kubenswrapper[4771]: E1002 11:48:37.896675 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" containerName="extract-content" Oct 02 11:48:37 crc kubenswrapper[4771]: I1002 11:48:37.896681 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" containerName="extract-content" Oct 02 11:48:37 crc kubenswrapper[4771]: E1002 11:48:37.896729 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" containerName="registry-server" Oct 02 11:48:37 crc kubenswrapper[4771]: I1002 11:48:37.896737 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" containerName="registry-server" Oct 02 11:48:37 crc kubenswrapper[4771]: I1002 11:48:37.896993 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dcad5d0-e29d-4df2-a8d0-eea71b69ebf8" containerName="registry-server" Oct 02 11:48:37 crc kubenswrapper[4771]: I1002 11:48:37.898764 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:37 crc kubenswrapper[4771]: I1002 11:48:37.919771 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2hj8n"] Oct 02 11:48:38 crc kubenswrapper[4771]: I1002 11:48:38.069854 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x74rx\" (UniqueName: \"kubernetes.io/projected/b8bd8af4-b12c-4325-9f32-c393c038a63b-kube-api-access-x74rx\") pod \"community-operators-2hj8n\" (UID: \"b8bd8af4-b12c-4325-9f32-c393c038a63b\") " pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:38 crc kubenswrapper[4771]: I1002 11:48:38.070403 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8bd8af4-b12c-4325-9f32-c393c038a63b-catalog-content\") pod \"community-operators-2hj8n\" (UID: \"b8bd8af4-b12c-4325-9f32-c393c038a63b\") " pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:38 crc kubenswrapper[4771]: I1002 11:48:38.070578 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8bd8af4-b12c-4325-9f32-c393c038a63b-utilities\") pod \"community-operators-2hj8n\" (UID: \"b8bd8af4-b12c-4325-9f32-c393c038a63b\") " pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:38 crc kubenswrapper[4771]: I1002 11:48:38.172548 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8bd8af4-b12c-4325-9f32-c393c038a63b-utilities\") pod \"community-operators-2hj8n\" (UID: \"b8bd8af4-b12c-4325-9f32-c393c038a63b\") " pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:38 crc kubenswrapper[4771]: I1002 11:48:38.172662 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x74rx\" (UniqueName: \"kubernetes.io/projected/b8bd8af4-b12c-4325-9f32-c393c038a63b-kube-api-access-x74rx\") pod \"community-operators-2hj8n\" (UID: \"b8bd8af4-b12c-4325-9f32-c393c038a63b\") " pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:38 crc kubenswrapper[4771]: I1002 11:48:38.173371 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8bd8af4-b12c-4325-9f32-c393c038a63b-utilities\") pod \"community-operators-2hj8n\" (UID: \"b8bd8af4-b12c-4325-9f32-c393c038a63b\") " pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:38 crc kubenswrapper[4771]: I1002 11:48:38.174351 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8bd8af4-b12c-4325-9f32-c393c038a63b-catalog-content\") pod \"community-operators-2hj8n\" (UID: \"b8bd8af4-b12c-4325-9f32-c393c038a63b\") " pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:38 crc kubenswrapper[4771]: I1002 11:48:38.173890 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8bd8af4-b12c-4325-9f32-c393c038a63b-catalog-content\") pod \"community-operators-2hj8n\" (UID: \"b8bd8af4-b12c-4325-9f32-c393c038a63b\") " pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:38 crc kubenswrapper[4771]: I1002 11:48:38.194108 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x74rx\" (UniqueName: \"kubernetes.io/projected/b8bd8af4-b12c-4325-9f32-c393c038a63b-kube-api-access-x74rx\") pod \"community-operators-2hj8n\" (UID: \"b8bd8af4-b12c-4325-9f32-c393c038a63b\") " pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:38 crc kubenswrapper[4771]: I1002 11:48:38.221553 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:38 crc kubenswrapper[4771]: I1002 11:48:38.786625 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2hj8n"] Oct 02 11:48:39 crc kubenswrapper[4771]: I1002 11:48:39.581165 4771 generic.go:334] "Generic (PLEG): container finished" podID="b8bd8af4-b12c-4325-9f32-c393c038a63b" containerID="231ad9b473b0736f6bc8ed45744a1dba95820f6cdd437d4e213f5e79774d2c7d" exitCode=0 Oct 02 11:48:39 crc kubenswrapper[4771]: I1002 11:48:39.581257 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2hj8n" event={"ID":"b8bd8af4-b12c-4325-9f32-c393c038a63b","Type":"ContainerDied","Data":"231ad9b473b0736f6bc8ed45744a1dba95820f6cdd437d4e213f5e79774d2c7d"} Oct 02 11:48:39 crc kubenswrapper[4771]: I1002 11:48:39.581531 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2hj8n" event={"ID":"b8bd8af4-b12c-4325-9f32-c393c038a63b","Type":"ContainerStarted","Data":"bb02df8ee167eefb5c3dd424ad6f2440d551ba23784133772c95bebf785f5774"} Oct 02 11:48:41 crc kubenswrapper[4771]: I1002 11:48:41.606300 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2hj8n" event={"ID":"b8bd8af4-b12c-4325-9f32-c393c038a63b","Type":"ContainerStarted","Data":"126e45a1b5e67ae74ae4b9ca3d1a4f8c95f0cd1cdda267b266bc7f9123a92f57"} Oct 02 11:48:42 crc kubenswrapper[4771]: I1002 11:48:42.618587 4771 generic.go:334] "Generic (PLEG): container finished" podID="b8bd8af4-b12c-4325-9f32-c393c038a63b" containerID="126e45a1b5e67ae74ae4b9ca3d1a4f8c95f0cd1cdda267b266bc7f9123a92f57" exitCode=0 Oct 02 11:48:42 crc kubenswrapper[4771]: I1002 11:48:42.618692 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2hj8n" event={"ID":"b8bd8af4-b12c-4325-9f32-c393c038a63b","Type":"ContainerDied","Data":"126e45a1b5e67ae74ae4b9ca3d1a4f8c95f0cd1cdda267b266bc7f9123a92f57"} Oct 02 11:48:44 crc kubenswrapper[4771]: I1002 11:48:44.646418 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2hj8n" event={"ID":"b8bd8af4-b12c-4325-9f32-c393c038a63b","Type":"ContainerStarted","Data":"3013dda9e799ccc2eb5f6e49416b6b2dee709031c4789549e5decc09b55e5c8a"} Oct 02 11:48:44 crc kubenswrapper[4771]: I1002 11:48:44.671496 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2hj8n" podStartSLOduration=3.690381242 podStartE2EDuration="7.671475905s" podCreationTimestamp="2025-10-02 11:48:37 +0000 UTC" firstStartedPulling="2025-10-02 11:48:39.584794286 +0000 UTC m=+7907.232479363" lastFinishedPulling="2025-10-02 11:48:43.565888959 +0000 UTC m=+7911.213574026" observedRunningTime="2025-10-02 11:48:44.66702033 +0000 UTC m=+7912.314705417" watchObservedRunningTime="2025-10-02 11:48:44.671475905 +0000 UTC m=+7912.319160972" Oct 02 11:48:48 crc kubenswrapper[4771]: I1002 11:48:48.222780 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:48 crc kubenswrapper[4771]: I1002 11:48:48.223335 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:48 crc kubenswrapper[4771]: I1002 11:48:48.280586 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:48 crc kubenswrapper[4771]: I1002 11:48:48.682389 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:48:48 crc kubenswrapper[4771]: E1002 11:48:48.682717 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:48:48 crc kubenswrapper[4771]: I1002 11:48:48.738937 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:48 crc kubenswrapper[4771]: I1002 11:48:48.790649 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2hj8n"] Oct 02 11:48:50 crc kubenswrapper[4771]: I1002 11:48:50.710234 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2hj8n" podUID="b8bd8af4-b12c-4325-9f32-c393c038a63b" containerName="registry-server" containerID="cri-o://3013dda9e799ccc2eb5f6e49416b6b2dee709031c4789549e5decc09b55e5c8a" gracePeriod=2 Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.231796 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.343546 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8bd8af4-b12c-4325-9f32-c393c038a63b-utilities\") pod \"b8bd8af4-b12c-4325-9f32-c393c038a63b\" (UID: \"b8bd8af4-b12c-4325-9f32-c393c038a63b\") " Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.344279 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8bd8af4-b12c-4325-9f32-c393c038a63b-utilities" (OuterVolumeSpecName: "utilities") pod "b8bd8af4-b12c-4325-9f32-c393c038a63b" (UID: "b8bd8af4-b12c-4325-9f32-c393c038a63b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.344385 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x74rx\" (UniqueName: \"kubernetes.io/projected/b8bd8af4-b12c-4325-9f32-c393c038a63b-kube-api-access-x74rx\") pod \"b8bd8af4-b12c-4325-9f32-c393c038a63b\" (UID: \"b8bd8af4-b12c-4325-9f32-c393c038a63b\") " Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.345266 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8bd8af4-b12c-4325-9f32-c393c038a63b-catalog-content\") pod \"b8bd8af4-b12c-4325-9f32-c393c038a63b\" (UID: \"b8bd8af4-b12c-4325-9f32-c393c038a63b\") " Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.345735 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8bd8af4-b12c-4325-9f32-c393c038a63b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.350718 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8bd8af4-b12c-4325-9f32-c393c038a63b-kube-api-access-x74rx" (OuterVolumeSpecName: "kube-api-access-x74rx") pod "b8bd8af4-b12c-4325-9f32-c393c038a63b" (UID: "b8bd8af4-b12c-4325-9f32-c393c038a63b"). InnerVolumeSpecName "kube-api-access-x74rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.397357 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8bd8af4-b12c-4325-9f32-c393c038a63b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b8bd8af4-b12c-4325-9f32-c393c038a63b" (UID: "b8bd8af4-b12c-4325-9f32-c393c038a63b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.446694 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x74rx\" (UniqueName: \"kubernetes.io/projected/b8bd8af4-b12c-4325-9f32-c393c038a63b-kube-api-access-x74rx\") on node \"crc\" DevicePath \"\"" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.446737 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8bd8af4-b12c-4325-9f32-c393c038a63b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.727810 4771 generic.go:334] "Generic (PLEG): container finished" podID="b8bd8af4-b12c-4325-9f32-c393c038a63b" containerID="3013dda9e799ccc2eb5f6e49416b6b2dee709031c4789549e5decc09b55e5c8a" exitCode=0 Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.727892 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2hj8n" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.727902 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2hj8n" event={"ID":"b8bd8af4-b12c-4325-9f32-c393c038a63b","Type":"ContainerDied","Data":"3013dda9e799ccc2eb5f6e49416b6b2dee709031c4789549e5decc09b55e5c8a"} Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.728423 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2hj8n" event={"ID":"b8bd8af4-b12c-4325-9f32-c393c038a63b","Type":"ContainerDied","Data":"bb02df8ee167eefb5c3dd424ad6f2440d551ba23784133772c95bebf785f5774"} Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.728443 4771 scope.go:117] "RemoveContainer" containerID="3013dda9e799ccc2eb5f6e49416b6b2dee709031c4789549e5decc09b55e5c8a" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.758979 4771 scope.go:117] "RemoveContainer" containerID="126e45a1b5e67ae74ae4b9ca3d1a4f8c95f0cd1cdda267b266bc7f9123a92f57" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.759605 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2hj8n"] Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.770760 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2hj8n"] Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.782500 4771 scope.go:117] "RemoveContainer" containerID="231ad9b473b0736f6bc8ed45744a1dba95820f6cdd437d4e213f5e79774d2c7d" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.847157 4771 scope.go:117] "RemoveContainer" containerID="3013dda9e799ccc2eb5f6e49416b6b2dee709031c4789549e5decc09b55e5c8a" Oct 02 11:48:51 crc kubenswrapper[4771]: E1002 11:48:51.847642 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3013dda9e799ccc2eb5f6e49416b6b2dee709031c4789549e5decc09b55e5c8a\": container with ID starting with 3013dda9e799ccc2eb5f6e49416b6b2dee709031c4789549e5decc09b55e5c8a not found: ID does not exist" containerID="3013dda9e799ccc2eb5f6e49416b6b2dee709031c4789549e5decc09b55e5c8a" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.847702 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3013dda9e799ccc2eb5f6e49416b6b2dee709031c4789549e5decc09b55e5c8a"} err="failed to get container status \"3013dda9e799ccc2eb5f6e49416b6b2dee709031c4789549e5decc09b55e5c8a\": rpc error: code = NotFound desc = could not find container \"3013dda9e799ccc2eb5f6e49416b6b2dee709031c4789549e5decc09b55e5c8a\": container with ID starting with 3013dda9e799ccc2eb5f6e49416b6b2dee709031c4789549e5decc09b55e5c8a not found: ID does not exist" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.847735 4771 scope.go:117] "RemoveContainer" containerID="126e45a1b5e67ae74ae4b9ca3d1a4f8c95f0cd1cdda267b266bc7f9123a92f57" Oct 02 11:48:51 crc kubenswrapper[4771]: E1002 11:48:51.848054 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"126e45a1b5e67ae74ae4b9ca3d1a4f8c95f0cd1cdda267b266bc7f9123a92f57\": container with ID starting with 126e45a1b5e67ae74ae4b9ca3d1a4f8c95f0cd1cdda267b266bc7f9123a92f57 not found: ID does not exist" containerID="126e45a1b5e67ae74ae4b9ca3d1a4f8c95f0cd1cdda267b266bc7f9123a92f57" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.848089 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"126e45a1b5e67ae74ae4b9ca3d1a4f8c95f0cd1cdda267b266bc7f9123a92f57"} err="failed to get container status \"126e45a1b5e67ae74ae4b9ca3d1a4f8c95f0cd1cdda267b266bc7f9123a92f57\": rpc error: code = NotFound desc = could not find container \"126e45a1b5e67ae74ae4b9ca3d1a4f8c95f0cd1cdda267b266bc7f9123a92f57\": container with ID starting with 126e45a1b5e67ae74ae4b9ca3d1a4f8c95f0cd1cdda267b266bc7f9123a92f57 not found: ID does not exist" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.848111 4771 scope.go:117] "RemoveContainer" containerID="231ad9b473b0736f6bc8ed45744a1dba95820f6cdd437d4e213f5e79774d2c7d" Oct 02 11:48:51 crc kubenswrapper[4771]: E1002 11:48:51.848372 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"231ad9b473b0736f6bc8ed45744a1dba95820f6cdd437d4e213f5e79774d2c7d\": container with ID starting with 231ad9b473b0736f6bc8ed45744a1dba95820f6cdd437d4e213f5e79774d2c7d not found: ID does not exist" containerID="231ad9b473b0736f6bc8ed45744a1dba95820f6cdd437d4e213f5e79774d2c7d" Oct 02 11:48:51 crc kubenswrapper[4771]: I1002 11:48:51.848405 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"231ad9b473b0736f6bc8ed45744a1dba95820f6cdd437d4e213f5e79774d2c7d"} err="failed to get container status \"231ad9b473b0736f6bc8ed45744a1dba95820f6cdd437d4e213f5e79774d2c7d\": rpc error: code = NotFound desc = could not find container \"231ad9b473b0736f6bc8ed45744a1dba95820f6cdd437d4e213f5e79774d2c7d\": container with ID starting with 231ad9b473b0736f6bc8ed45744a1dba95820f6cdd437d4e213f5e79774d2c7d not found: ID does not exist" Oct 02 11:48:53 crc kubenswrapper[4771]: I1002 11:48:53.701892 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8bd8af4-b12c-4325-9f32-c393c038a63b" path="/var/lib/kubelet/pods/b8bd8af4-b12c-4325-9f32-c393c038a63b/volumes" Oct 02 11:49:02 crc kubenswrapper[4771]: I1002 11:49:02.681994 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:49:02 crc kubenswrapper[4771]: E1002 11:49:02.682941 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:49:17 crc kubenswrapper[4771]: I1002 11:49:17.681292 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:49:17 crc kubenswrapper[4771]: E1002 11:49:17.683728 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:49:32 crc kubenswrapper[4771]: I1002 11:49:32.682035 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:49:32 crc kubenswrapper[4771]: E1002 11:49:32.683562 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:49:45 crc kubenswrapper[4771]: I1002 11:49:45.682016 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:49:45 crc kubenswrapper[4771]: E1002 11:49:45.682869 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:49:57 crc kubenswrapper[4771]: I1002 11:49:57.682288 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:49:57 crc kubenswrapper[4771]: E1002 11:49:57.683310 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:50:12 crc kubenswrapper[4771]: I1002 11:50:12.681354 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:50:12 crc kubenswrapper[4771]: E1002 11:50:12.682177 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:50:26 crc kubenswrapper[4771]: I1002 11:50:26.682723 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:50:26 crc kubenswrapper[4771]: E1002 11:50:26.683603 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:50:37 crc kubenswrapper[4771]: I1002 11:50:37.681707 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:50:37 crc kubenswrapper[4771]: E1002 11:50:37.682592 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:50:48 crc kubenswrapper[4771]: I1002 11:50:48.681826 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:50:48 crc kubenswrapper[4771]: E1002 11:50:48.682736 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:51:02 crc kubenswrapper[4771]: I1002 11:51:02.681000 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:51:02 crc kubenswrapper[4771]: E1002 11:51:02.681748 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:51:08 crc kubenswrapper[4771]: I1002 11:51:08.722929 4771 scope.go:117] "RemoveContainer" containerID="3bb7ea0a141a602949a8bd1148c94c21c4548c1d184717a4655bd3bbcfb4afeb" Oct 02 11:51:15 crc kubenswrapper[4771]: I1002 11:51:15.681832 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:51:15 crc kubenswrapper[4771]: E1002 11:51:15.682774 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:51:27 crc kubenswrapper[4771]: I1002 11:51:27.682102 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:51:27 crc kubenswrapper[4771]: E1002 11:51:27.682831 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:51:39 crc kubenswrapper[4771]: I1002 11:51:39.683064 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:51:39 crc kubenswrapper[4771]: E1002 11:51:39.686471 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:51:50 crc kubenswrapper[4771]: I1002 11:51:50.682243 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:51:51 crc kubenswrapper[4771]: I1002 11:51:51.751590 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"79a9cdd95ae64294ca9c39ad08d44cdc8883948e1e7f87ca95d709f164e260ab"} Oct 02 11:54:12 crc kubenswrapper[4771]: I1002 11:54:12.146860 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:54:12 crc kubenswrapper[4771]: I1002 11:54:12.147703 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.231081 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t65fn"] Oct 02 11:54:23 crc kubenswrapper[4771]: E1002 11:54:23.232478 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8bd8af4-b12c-4325-9f32-c393c038a63b" containerName="extract-utilities" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.232499 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8bd8af4-b12c-4325-9f32-c393c038a63b" containerName="extract-utilities" Oct 02 11:54:23 crc kubenswrapper[4771]: E1002 11:54:23.232550 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8bd8af4-b12c-4325-9f32-c393c038a63b" containerName="extract-content" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.232559 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8bd8af4-b12c-4325-9f32-c393c038a63b" containerName="extract-content" Oct 02 11:54:23 crc kubenswrapper[4771]: E1002 11:54:23.232604 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8bd8af4-b12c-4325-9f32-c393c038a63b" containerName="registry-server" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.232613 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8bd8af4-b12c-4325-9f32-c393c038a63b" containerName="registry-server" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.232884 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8bd8af4-b12c-4325-9f32-c393c038a63b" containerName="registry-server" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.235114 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.246366 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t65fn"] Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.359292 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-utilities\") pod \"certified-operators-t65fn\" (UID: \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\") " pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.360292 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-catalog-content\") pod \"certified-operators-t65fn\" (UID: \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\") " pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.360395 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn6mn\" (UniqueName: \"kubernetes.io/projected/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-kube-api-access-tn6mn\") pod \"certified-operators-t65fn\" (UID: \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\") " pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.463047 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-utilities\") pod \"certified-operators-t65fn\" (UID: \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\") " pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.463575 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-catalog-content\") pod \"certified-operators-t65fn\" (UID: \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\") " pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.463602 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn6mn\" (UniqueName: \"kubernetes.io/projected/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-kube-api-access-tn6mn\") pod \"certified-operators-t65fn\" (UID: \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\") " pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.467255 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-utilities\") pod \"certified-operators-t65fn\" (UID: \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\") " pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.467964 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-catalog-content\") pod \"certified-operators-t65fn\" (UID: \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\") " pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.497096 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn6mn\" (UniqueName: \"kubernetes.io/projected/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-kube-api-access-tn6mn\") pod \"certified-operators-t65fn\" (UID: \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\") " pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:23 crc kubenswrapper[4771]: I1002 11:54:23.565721 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:24 crc kubenswrapper[4771]: I1002 11:54:24.244233 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t65fn"] Oct 02 11:54:24 crc kubenswrapper[4771]: I1002 11:54:24.659925 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t65fn" event={"ID":"ead4b747-61c3-4cc5-8fc6-12313efb0ea4","Type":"ContainerStarted","Data":"3d8f5b80bf505940dfe5b94d0cbf3b62255abaef8e8c6fb02e2f62a888178104"} Oct 02 11:54:25 crc kubenswrapper[4771]: I1002 11:54:25.671915 4771 generic.go:334] "Generic (PLEG): container finished" podID="ead4b747-61c3-4cc5-8fc6-12313efb0ea4" containerID="ae9f88c143ab17f7708ddcc4e4a363886f8ad72547cf2cbae07f018ff4851b68" exitCode=0 Oct 02 11:54:25 crc kubenswrapper[4771]: I1002 11:54:25.671991 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t65fn" event={"ID":"ead4b747-61c3-4cc5-8fc6-12313efb0ea4","Type":"ContainerDied","Data":"ae9f88c143ab17f7708ddcc4e4a363886f8ad72547cf2cbae07f018ff4851b68"} Oct 02 11:54:25 crc kubenswrapper[4771]: I1002 11:54:25.675260 4771 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:54:26 crc kubenswrapper[4771]: I1002 11:54:26.683037 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t65fn" event={"ID":"ead4b747-61c3-4cc5-8fc6-12313efb0ea4","Type":"ContainerStarted","Data":"351b376e03ca59eb804c1d528404eacbb640a625f0690b85af386f3d48fff9be"} Oct 02 11:54:27 crc kubenswrapper[4771]: I1002 11:54:27.227515 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ghblb"] Oct 02 11:54:27 crc kubenswrapper[4771]: I1002 11:54:27.230817 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:27 crc kubenswrapper[4771]: I1002 11:54:27.238601 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghblb"] Oct 02 11:54:27 crc kubenswrapper[4771]: I1002 11:54:27.274626 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qzm7\" (UniqueName: \"kubernetes.io/projected/e395ad53-db6d-4495-9bde-13e6397d4c5b-kube-api-access-2qzm7\") pod \"redhat-marketplace-ghblb\" (UID: \"e395ad53-db6d-4495-9bde-13e6397d4c5b\") " pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:27 crc kubenswrapper[4771]: I1002 11:54:27.275015 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e395ad53-db6d-4495-9bde-13e6397d4c5b-utilities\") pod \"redhat-marketplace-ghblb\" (UID: \"e395ad53-db6d-4495-9bde-13e6397d4c5b\") " pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:27 crc kubenswrapper[4771]: I1002 11:54:27.275414 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e395ad53-db6d-4495-9bde-13e6397d4c5b-catalog-content\") pod \"redhat-marketplace-ghblb\" (UID: \"e395ad53-db6d-4495-9bde-13e6397d4c5b\") " pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:27 crc kubenswrapper[4771]: I1002 11:54:27.378184 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qzm7\" (UniqueName: \"kubernetes.io/projected/e395ad53-db6d-4495-9bde-13e6397d4c5b-kube-api-access-2qzm7\") pod \"redhat-marketplace-ghblb\" (UID: \"e395ad53-db6d-4495-9bde-13e6397d4c5b\") " pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:27 crc kubenswrapper[4771]: I1002 11:54:27.378744 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e395ad53-db6d-4495-9bde-13e6397d4c5b-utilities\") pod \"redhat-marketplace-ghblb\" (UID: \"e395ad53-db6d-4495-9bde-13e6397d4c5b\") " pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:27 crc kubenswrapper[4771]: I1002 11:54:27.379257 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e395ad53-db6d-4495-9bde-13e6397d4c5b-catalog-content\") pod \"redhat-marketplace-ghblb\" (UID: \"e395ad53-db6d-4495-9bde-13e6397d4c5b\") " pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:27 crc kubenswrapper[4771]: I1002 11:54:27.398459 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e395ad53-db6d-4495-9bde-13e6397d4c5b-utilities\") pod \"redhat-marketplace-ghblb\" (UID: \"e395ad53-db6d-4495-9bde-13e6397d4c5b\") " pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:27 crc kubenswrapper[4771]: I1002 11:54:27.398693 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e395ad53-db6d-4495-9bde-13e6397d4c5b-catalog-content\") pod \"redhat-marketplace-ghblb\" (UID: \"e395ad53-db6d-4495-9bde-13e6397d4c5b\") " pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:27 crc kubenswrapper[4771]: I1002 11:54:27.406462 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qzm7\" (UniqueName: \"kubernetes.io/projected/e395ad53-db6d-4495-9bde-13e6397d4c5b-kube-api-access-2qzm7\") pod \"redhat-marketplace-ghblb\" (UID: \"e395ad53-db6d-4495-9bde-13e6397d4c5b\") " pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:27 crc kubenswrapper[4771]: I1002 11:54:27.586867 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:28 crc kubenswrapper[4771]: I1002 11:54:28.198436 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghblb"] Oct 02 11:54:28 crc kubenswrapper[4771]: W1002 11:54:28.206848 4771 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode395ad53_db6d_4495_9bde_13e6397d4c5b.slice/crio-0270e65f2a28bcf05f2bc9f42179af75f125fb784ad8dffe97c3d3811614933a WatchSource:0}: Error finding container 0270e65f2a28bcf05f2bc9f42179af75f125fb784ad8dffe97c3d3811614933a: Status 404 returned error can't find the container with id 0270e65f2a28bcf05f2bc9f42179af75f125fb784ad8dffe97c3d3811614933a Oct 02 11:54:28 crc kubenswrapper[4771]: I1002 11:54:28.711239 4771 generic.go:334] "Generic (PLEG): container finished" podID="e395ad53-db6d-4495-9bde-13e6397d4c5b" containerID="6e2046e87e1075b48489c2cbf70336eb2c02fed25c020cc180f19f82c9b1347e" exitCode=0 Oct 02 11:54:28 crc kubenswrapper[4771]: I1002 11:54:28.711335 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghblb" event={"ID":"e395ad53-db6d-4495-9bde-13e6397d4c5b","Type":"ContainerDied","Data":"6e2046e87e1075b48489c2cbf70336eb2c02fed25c020cc180f19f82c9b1347e"} Oct 02 11:54:28 crc kubenswrapper[4771]: I1002 11:54:28.711651 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghblb" event={"ID":"e395ad53-db6d-4495-9bde-13e6397d4c5b","Type":"ContainerStarted","Data":"0270e65f2a28bcf05f2bc9f42179af75f125fb784ad8dffe97c3d3811614933a"} Oct 02 11:54:28 crc kubenswrapper[4771]: I1002 11:54:28.720963 4771 generic.go:334] "Generic (PLEG): container finished" podID="ead4b747-61c3-4cc5-8fc6-12313efb0ea4" containerID="351b376e03ca59eb804c1d528404eacbb640a625f0690b85af386f3d48fff9be" exitCode=0 Oct 02 11:54:28 crc kubenswrapper[4771]: I1002 11:54:28.721023 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t65fn" event={"ID":"ead4b747-61c3-4cc5-8fc6-12313efb0ea4","Type":"ContainerDied","Data":"351b376e03ca59eb804c1d528404eacbb640a625f0690b85af386f3d48fff9be"} Oct 02 11:54:29 crc kubenswrapper[4771]: I1002 11:54:29.740653 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghblb" event={"ID":"e395ad53-db6d-4495-9bde-13e6397d4c5b","Type":"ContainerStarted","Data":"1c73227fa900aff0953df922d14cd62aa00bfb7b689fd7faceeacd1f461e293b"} Oct 02 11:54:29 crc kubenswrapper[4771]: I1002 11:54:29.744659 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t65fn" event={"ID":"ead4b747-61c3-4cc5-8fc6-12313efb0ea4","Type":"ContainerStarted","Data":"867a4be2f412b1548f075d69a463e07acdcbd21f7a350335631783d4a187169a"} Oct 02 11:54:29 crc kubenswrapper[4771]: I1002 11:54:29.783077 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t65fn" podStartSLOduration=3.199150192 podStartE2EDuration="6.783048859s" podCreationTimestamp="2025-10-02 11:54:23 +0000 UTC" firstStartedPulling="2025-10-02 11:54:25.675021359 +0000 UTC m=+8253.322706426" lastFinishedPulling="2025-10-02 11:54:29.258920026 +0000 UTC m=+8256.906605093" observedRunningTime="2025-10-02 11:54:29.779350104 +0000 UTC m=+8257.427035181" watchObservedRunningTime="2025-10-02 11:54:29.783048859 +0000 UTC m=+8257.430733926" Oct 02 11:54:31 crc kubenswrapper[4771]: I1002 11:54:31.780503 4771 generic.go:334] "Generic (PLEG): container finished" podID="e395ad53-db6d-4495-9bde-13e6397d4c5b" containerID="1c73227fa900aff0953df922d14cd62aa00bfb7b689fd7faceeacd1f461e293b" exitCode=0 Oct 02 11:54:31 crc kubenswrapper[4771]: I1002 11:54:31.780633 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghblb" event={"ID":"e395ad53-db6d-4495-9bde-13e6397d4c5b","Type":"ContainerDied","Data":"1c73227fa900aff0953df922d14cd62aa00bfb7b689fd7faceeacd1f461e293b"} Oct 02 11:54:32 crc kubenswrapper[4771]: I1002 11:54:32.797879 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghblb" event={"ID":"e395ad53-db6d-4495-9bde-13e6397d4c5b","Type":"ContainerStarted","Data":"692e3f8e58ec123d9c468f9de950390b28ae337a8a89a251917d2cd0a8f3f5d4"} Oct 02 11:54:32 crc kubenswrapper[4771]: I1002 11:54:32.821921 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ghblb" podStartSLOduration=2.249658017 podStartE2EDuration="5.821901683s" podCreationTimestamp="2025-10-02 11:54:27 +0000 UTC" firstStartedPulling="2025-10-02 11:54:28.714706824 +0000 UTC m=+8256.362391891" lastFinishedPulling="2025-10-02 11:54:32.28695049 +0000 UTC m=+8259.934635557" observedRunningTime="2025-10-02 11:54:32.817084598 +0000 UTC m=+8260.464769665" watchObservedRunningTime="2025-10-02 11:54:32.821901683 +0000 UTC m=+8260.469586750" Oct 02 11:54:33 crc kubenswrapper[4771]: I1002 11:54:33.566908 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:33 crc kubenswrapper[4771]: I1002 11:54:33.566959 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:34 crc kubenswrapper[4771]: I1002 11:54:34.639463 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-t65fn" podUID="ead4b747-61c3-4cc5-8fc6-12313efb0ea4" containerName="registry-server" probeResult="failure" output=< Oct 02 11:54:34 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:54:34 crc kubenswrapper[4771]: > Oct 02 11:54:37 crc kubenswrapper[4771]: I1002 11:54:37.587295 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:37 crc kubenswrapper[4771]: I1002 11:54:37.587899 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:37 crc kubenswrapper[4771]: I1002 11:54:37.692521 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:37 crc kubenswrapper[4771]: I1002 11:54:37.930948 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:39 crc kubenswrapper[4771]: I1002 11:54:39.829073 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghblb"] Oct 02 11:54:39 crc kubenswrapper[4771]: I1002 11:54:39.886503 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ghblb" podUID="e395ad53-db6d-4495-9bde-13e6397d4c5b" containerName="registry-server" containerID="cri-o://692e3f8e58ec123d9c468f9de950390b28ae337a8a89a251917d2cd0a8f3f5d4" gracePeriod=2 Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.663926 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.679341 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e395ad53-db6d-4495-9bde-13e6397d4c5b-catalog-content\") pod \"e395ad53-db6d-4495-9bde-13e6397d4c5b\" (UID: \"e395ad53-db6d-4495-9bde-13e6397d4c5b\") " Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.679662 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e395ad53-db6d-4495-9bde-13e6397d4c5b-utilities\") pod \"e395ad53-db6d-4495-9bde-13e6397d4c5b\" (UID: \"e395ad53-db6d-4495-9bde-13e6397d4c5b\") " Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.679739 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qzm7\" (UniqueName: \"kubernetes.io/projected/e395ad53-db6d-4495-9bde-13e6397d4c5b-kube-api-access-2qzm7\") pod \"e395ad53-db6d-4495-9bde-13e6397d4c5b\" (UID: \"e395ad53-db6d-4495-9bde-13e6397d4c5b\") " Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.680645 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e395ad53-db6d-4495-9bde-13e6397d4c5b-utilities" (OuterVolumeSpecName: "utilities") pod "e395ad53-db6d-4495-9bde-13e6397d4c5b" (UID: "e395ad53-db6d-4495-9bde-13e6397d4c5b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.699397 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e395ad53-db6d-4495-9bde-13e6397d4c5b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e395ad53-db6d-4495-9bde-13e6397d4c5b" (UID: "e395ad53-db6d-4495-9bde-13e6397d4c5b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.709545 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e395ad53-db6d-4495-9bde-13e6397d4c5b-kube-api-access-2qzm7" (OuterVolumeSpecName: "kube-api-access-2qzm7") pod "e395ad53-db6d-4495-9bde-13e6397d4c5b" (UID: "e395ad53-db6d-4495-9bde-13e6397d4c5b"). InnerVolumeSpecName "kube-api-access-2qzm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.787952 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e395ad53-db6d-4495-9bde-13e6397d4c5b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.787995 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e395ad53-db6d-4495-9bde-13e6397d4c5b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.788007 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qzm7\" (UniqueName: \"kubernetes.io/projected/e395ad53-db6d-4495-9bde-13e6397d4c5b-kube-api-access-2qzm7\") on node \"crc\" DevicePath \"\"" Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.902482 4771 generic.go:334] "Generic (PLEG): container finished" podID="e395ad53-db6d-4495-9bde-13e6397d4c5b" containerID="692e3f8e58ec123d9c468f9de950390b28ae337a8a89a251917d2cd0a8f3f5d4" exitCode=0 Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.902562 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghblb" event={"ID":"e395ad53-db6d-4495-9bde-13e6397d4c5b","Type":"ContainerDied","Data":"692e3f8e58ec123d9c468f9de950390b28ae337a8a89a251917d2cd0a8f3f5d4"} Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.902622 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghblb" event={"ID":"e395ad53-db6d-4495-9bde-13e6397d4c5b","Type":"ContainerDied","Data":"0270e65f2a28bcf05f2bc9f42179af75f125fb784ad8dffe97c3d3811614933a"} Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.902658 4771 scope.go:117] "RemoveContainer" containerID="692e3f8e58ec123d9c468f9de950390b28ae337a8a89a251917d2cd0a8f3f5d4" Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.903945 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ghblb" Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.937560 4771 scope.go:117] "RemoveContainer" containerID="1c73227fa900aff0953df922d14cd62aa00bfb7b689fd7faceeacd1f461e293b" Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.948900 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghblb"] Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.965123 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghblb"] Oct 02 11:54:40 crc kubenswrapper[4771]: I1002 11:54:40.999118 4771 scope.go:117] "RemoveContainer" containerID="6e2046e87e1075b48489c2cbf70336eb2c02fed25c020cc180f19f82c9b1347e" Oct 02 11:54:41 crc kubenswrapper[4771]: I1002 11:54:41.047305 4771 scope.go:117] "RemoveContainer" containerID="692e3f8e58ec123d9c468f9de950390b28ae337a8a89a251917d2cd0a8f3f5d4" Oct 02 11:54:41 crc kubenswrapper[4771]: E1002 11:54:41.048472 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"692e3f8e58ec123d9c468f9de950390b28ae337a8a89a251917d2cd0a8f3f5d4\": container with ID starting with 692e3f8e58ec123d9c468f9de950390b28ae337a8a89a251917d2cd0a8f3f5d4 not found: ID does not exist" containerID="692e3f8e58ec123d9c468f9de950390b28ae337a8a89a251917d2cd0a8f3f5d4" Oct 02 11:54:41 crc kubenswrapper[4771]: I1002 11:54:41.048597 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"692e3f8e58ec123d9c468f9de950390b28ae337a8a89a251917d2cd0a8f3f5d4"} err="failed to get container status \"692e3f8e58ec123d9c468f9de950390b28ae337a8a89a251917d2cd0a8f3f5d4\": rpc error: code = NotFound desc = could not find container \"692e3f8e58ec123d9c468f9de950390b28ae337a8a89a251917d2cd0a8f3f5d4\": container with ID starting with 692e3f8e58ec123d9c468f9de950390b28ae337a8a89a251917d2cd0a8f3f5d4 not found: ID does not exist" Oct 02 11:54:41 crc kubenswrapper[4771]: I1002 11:54:41.048647 4771 scope.go:117] "RemoveContainer" containerID="1c73227fa900aff0953df922d14cd62aa00bfb7b689fd7faceeacd1f461e293b" Oct 02 11:54:41 crc kubenswrapper[4771]: E1002 11:54:41.049317 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c73227fa900aff0953df922d14cd62aa00bfb7b689fd7faceeacd1f461e293b\": container with ID starting with 1c73227fa900aff0953df922d14cd62aa00bfb7b689fd7faceeacd1f461e293b not found: ID does not exist" containerID="1c73227fa900aff0953df922d14cd62aa00bfb7b689fd7faceeacd1f461e293b" Oct 02 11:54:41 crc kubenswrapper[4771]: I1002 11:54:41.049436 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c73227fa900aff0953df922d14cd62aa00bfb7b689fd7faceeacd1f461e293b"} err="failed to get container status \"1c73227fa900aff0953df922d14cd62aa00bfb7b689fd7faceeacd1f461e293b\": rpc error: code = NotFound desc = could not find container \"1c73227fa900aff0953df922d14cd62aa00bfb7b689fd7faceeacd1f461e293b\": container with ID starting with 1c73227fa900aff0953df922d14cd62aa00bfb7b689fd7faceeacd1f461e293b not found: ID does not exist" Oct 02 11:54:41 crc kubenswrapper[4771]: I1002 11:54:41.049519 4771 scope.go:117] "RemoveContainer" containerID="6e2046e87e1075b48489c2cbf70336eb2c02fed25c020cc180f19f82c9b1347e" Oct 02 11:54:41 crc kubenswrapper[4771]: E1002 11:54:41.049976 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e2046e87e1075b48489c2cbf70336eb2c02fed25c020cc180f19f82c9b1347e\": container with ID starting with 6e2046e87e1075b48489c2cbf70336eb2c02fed25c020cc180f19f82c9b1347e not found: ID does not exist" containerID="6e2046e87e1075b48489c2cbf70336eb2c02fed25c020cc180f19f82c9b1347e" Oct 02 11:54:41 crc kubenswrapper[4771]: I1002 11:54:41.050066 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e2046e87e1075b48489c2cbf70336eb2c02fed25c020cc180f19f82c9b1347e"} err="failed to get container status \"6e2046e87e1075b48489c2cbf70336eb2c02fed25c020cc180f19f82c9b1347e\": rpc error: code = NotFound desc = could not find container \"6e2046e87e1075b48489c2cbf70336eb2c02fed25c020cc180f19f82c9b1347e\": container with ID starting with 6e2046e87e1075b48489c2cbf70336eb2c02fed25c020cc180f19f82c9b1347e not found: ID does not exist" Oct 02 11:54:41 crc kubenswrapper[4771]: I1002 11:54:41.707090 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e395ad53-db6d-4495-9bde-13e6397d4c5b" path="/var/lib/kubelet/pods/e395ad53-db6d-4495-9bde-13e6397d4c5b/volumes" Oct 02 11:54:42 crc kubenswrapper[4771]: I1002 11:54:42.146275 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:54:42 crc kubenswrapper[4771]: I1002 11:54:42.146377 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:54:43 crc kubenswrapper[4771]: I1002 11:54:43.623661 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:43 crc kubenswrapper[4771]: I1002 11:54:43.705782 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:44 crc kubenswrapper[4771]: I1002 11:54:44.025864 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t65fn"] Oct 02 11:54:44 crc kubenswrapper[4771]: I1002 11:54:44.958157 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t65fn" podUID="ead4b747-61c3-4cc5-8fc6-12313efb0ea4" containerName="registry-server" containerID="cri-o://867a4be2f412b1548f075d69a463e07acdcbd21f7a350335631783d4a187169a" gracePeriod=2 Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.586677 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.766849 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tn6mn\" (UniqueName: \"kubernetes.io/projected/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-kube-api-access-tn6mn\") pod \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\" (UID: \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\") " Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.767865 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-catalog-content\") pod \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\" (UID: \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\") " Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.767983 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-utilities\") pod \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\" (UID: \"ead4b747-61c3-4cc5-8fc6-12313efb0ea4\") " Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.769656 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-utilities" (OuterVolumeSpecName: "utilities") pod "ead4b747-61c3-4cc5-8fc6-12313efb0ea4" (UID: "ead4b747-61c3-4cc5-8fc6-12313efb0ea4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.785938 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-kube-api-access-tn6mn" (OuterVolumeSpecName: "kube-api-access-tn6mn") pod "ead4b747-61c3-4cc5-8fc6-12313efb0ea4" (UID: "ead4b747-61c3-4cc5-8fc6-12313efb0ea4"). InnerVolumeSpecName "kube-api-access-tn6mn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.835854 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ead4b747-61c3-4cc5-8fc6-12313efb0ea4" (UID: "ead4b747-61c3-4cc5-8fc6-12313efb0ea4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.872724 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tn6mn\" (UniqueName: \"kubernetes.io/projected/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-kube-api-access-tn6mn\") on node \"crc\" DevicePath \"\"" Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.872766 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.872780 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ead4b747-61c3-4cc5-8fc6-12313efb0ea4-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.974643 4771 generic.go:334] "Generic (PLEG): container finished" podID="ead4b747-61c3-4cc5-8fc6-12313efb0ea4" containerID="867a4be2f412b1548f075d69a463e07acdcbd21f7a350335631783d4a187169a" exitCode=0 Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.974701 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t65fn" Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.974708 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t65fn" event={"ID":"ead4b747-61c3-4cc5-8fc6-12313efb0ea4","Type":"ContainerDied","Data":"867a4be2f412b1548f075d69a463e07acdcbd21f7a350335631783d4a187169a"} Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.974787 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t65fn" event={"ID":"ead4b747-61c3-4cc5-8fc6-12313efb0ea4","Type":"ContainerDied","Data":"3d8f5b80bf505940dfe5b94d0cbf3b62255abaef8e8c6fb02e2f62a888178104"} Oct 02 11:54:45 crc kubenswrapper[4771]: I1002 11:54:45.974815 4771 scope.go:117] "RemoveContainer" containerID="867a4be2f412b1548f075d69a463e07acdcbd21f7a350335631783d4a187169a" Oct 02 11:54:46 crc kubenswrapper[4771]: I1002 11:54:46.033233 4771 scope.go:117] "RemoveContainer" containerID="351b376e03ca59eb804c1d528404eacbb640a625f0690b85af386f3d48fff9be" Oct 02 11:54:46 crc kubenswrapper[4771]: I1002 11:54:46.034194 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t65fn"] Oct 02 11:54:46 crc kubenswrapper[4771]: I1002 11:54:46.078664 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t65fn"] Oct 02 11:54:46 crc kubenswrapper[4771]: I1002 11:54:46.086084 4771 scope.go:117] "RemoveContainer" containerID="ae9f88c143ab17f7708ddcc4e4a363886f8ad72547cf2cbae07f018ff4851b68" Oct 02 11:54:46 crc kubenswrapper[4771]: I1002 11:54:46.153871 4771 scope.go:117] "RemoveContainer" containerID="867a4be2f412b1548f075d69a463e07acdcbd21f7a350335631783d4a187169a" Oct 02 11:54:46 crc kubenswrapper[4771]: E1002 11:54:46.154633 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"867a4be2f412b1548f075d69a463e07acdcbd21f7a350335631783d4a187169a\": container with ID starting with 867a4be2f412b1548f075d69a463e07acdcbd21f7a350335631783d4a187169a not found: ID does not exist" containerID="867a4be2f412b1548f075d69a463e07acdcbd21f7a350335631783d4a187169a" Oct 02 11:54:46 crc kubenswrapper[4771]: I1002 11:54:46.154688 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"867a4be2f412b1548f075d69a463e07acdcbd21f7a350335631783d4a187169a"} err="failed to get container status \"867a4be2f412b1548f075d69a463e07acdcbd21f7a350335631783d4a187169a\": rpc error: code = NotFound desc = could not find container \"867a4be2f412b1548f075d69a463e07acdcbd21f7a350335631783d4a187169a\": container with ID starting with 867a4be2f412b1548f075d69a463e07acdcbd21f7a350335631783d4a187169a not found: ID does not exist" Oct 02 11:54:46 crc kubenswrapper[4771]: I1002 11:54:46.154725 4771 scope.go:117] "RemoveContainer" containerID="351b376e03ca59eb804c1d528404eacbb640a625f0690b85af386f3d48fff9be" Oct 02 11:54:46 crc kubenswrapper[4771]: E1002 11:54:46.155110 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"351b376e03ca59eb804c1d528404eacbb640a625f0690b85af386f3d48fff9be\": container with ID starting with 351b376e03ca59eb804c1d528404eacbb640a625f0690b85af386f3d48fff9be not found: ID does not exist" containerID="351b376e03ca59eb804c1d528404eacbb640a625f0690b85af386f3d48fff9be" Oct 02 11:54:46 crc kubenswrapper[4771]: I1002 11:54:46.155148 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"351b376e03ca59eb804c1d528404eacbb640a625f0690b85af386f3d48fff9be"} err="failed to get container status \"351b376e03ca59eb804c1d528404eacbb640a625f0690b85af386f3d48fff9be\": rpc error: code = NotFound desc = could not find container \"351b376e03ca59eb804c1d528404eacbb640a625f0690b85af386f3d48fff9be\": container with ID starting with 351b376e03ca59eb804c1d528404eacbb640a625f0690b85af386f3d48fff9be not found: ID does not exist" Oct 02 11:54:46 crc kubenswrapper[4771]: I1002 11:54:46.155162 4771 scope.go:117] "RemoveContainer" containerID="ae9f88c143ab17f7708ddcc4e4a363886f8ad72547cf2cbae07f018ff4851b68" Oct 02 11:54:46 crc kubenswrapper[4771]: E1002 11:54:46.155401 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae9f88c143ab17f7708ddcc4e4a363886f8ad72547cf2cbae07f018ff4851b68\": container with ID starting with ae9f88c143ab17f7708ddcc4e4a363886f8ad72547cf2cbae07f018ff4851b68 not found: ID does not exist" containerID="ae9f88c143ab17f7708ddcc4e4a363886f8ad72547cf2cbae07f018ff4851b68" Oct 02 11:54:46 crc kubenswrapper[4771]: I1002 11:54:46.155431 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae9f88c143ab17f7708ddcc4e4a363886f8ad72547cf2cbae07f018ff4851b68"} err="failed to get container status \"ae9f88c143ab17f7708ddcc4e4a363886f8ad72547cf2cbae07f018ff4851b68\": rpc error: code = NotFound desc = could not find container \"ae9f88c143ab17f7708ddcc4e4a363886f8ad72547cf2cbae07f018ff4851b68\": container with ID starting with ae9f88c143ab17f7708ddcc4e4a363886f8ad72547cf2cbae07f018ff4851b68 not found: ID does not exist" Oct 02 11:54:47 crc kubenswrapper[4771]: I1002 11:54:47.701269 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ead4b747-61c3-4cc5-8fc6-12313efb0ea4" path="/var/lib/kubelet/pods/ead4b747-61c3-4cc5-8fc6-12313efb0ea4/volumes" Oct 02 11:55:12 crc kubenswrapper[4771]: I1002 11:55:12.146358 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:55:12 crc kubenswrapper[4771]: I1002 11:55:12.146948 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:55:12 crc kubenswrapper[4771]: I1002 11:55:12.147004 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 11:55:12 crc kubenswrapper[4771]: I1002 11:55:12.147930 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"79a9cdd95ae64294ca9c39ad08d44cdc8883948e1e7f87ca95d709f164e260ab"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:55:12 crc kubenswrapper[4771]: I1002 11:55:12.148000 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://79a9cdd95ae64294ca9c39ad08d44cdc8883948e1e7f87ca95d709f164e260ab" gracePeriod=600 Oct 02 11:55:12 crc kubenswrapper[4771]: I1002 11:55:12.330827 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="79a9cdd95ae64294ca9c39ad08d44cdc8883948e1e7f87ca95d709f164e260ab" exitCode=0 Oct 02 11:55:12 crc kubenswrapper[4771]: I1002 11:55:12.330929 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"79a9cdd95ae64294ca9c39ad08d44cdc8883948e1e7f87ca95d709f164e260ab"} Oct 02 11:55:12 crc kubenswrapper[4771]: I1002 11:55:12.331252 4771 scope.go:117] "RemoveContainer" containerID="bb55682929ba7460af728106cbff2f5017111d44278ae4ac9ada0f00423a40c5" Oct 02 11:55:13 crc kubenswrapper[4771]: I1002 11:55:13.347673 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096"} Oct 02 11:57:12 crc kubenswrapper[4771]: I1002 11:57:12.146299 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:57:12 crc kubenswrapper[4771]: I1002 11:57:12.147190 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:57:42 crc kubenswrapper[4771]: I1002 11:57:42.145832 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:57:42 crc kubenswrapper[4771]: I1002 11:57:42.146437 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:58:12 crc kubenswrapper[4771]: I1002 11:58:12.150567 4771 patch_prober.go:28] interesting pod/machine-config-daemon-m4mpt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:58:12 crc kubenswrapper[4771]: I1002 11:58:12.151779 4771 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:58:12 crc kubenswrapper[4771]: I1002 11:58:12.151981 4771 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" Oct 02 11:58:12 crc kubenswrapper[4771]: I1002 11:58:12.153326 4771 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096"} pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:58:12 crc kubenswrapper[4771]: I1002 11:58:12.153404 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerName="machine-config-daemon" containerID="cri-o://1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" gracePeriod=600 Oct 02 11:58:12 crc kubenswrapper[4771]: E1002 11:58:12.291265 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:58:12 crc kubenswrapper[4771]: I1002 11:58:12.688826 4771 generic.go:334] "Generic (PLEG): container finished" podID="52d686b1-d29a-42c4-97ce-e239a0c680b0" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" exitCode=0 Oct 02 11:58:12 crc kubenswrapper[4771]: I1002 11:58:12.688931 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerDied","Data":"1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096"} Oct 02 11:58:12 crc kubenswrapper[4771]: I1002 11:58:12.689421 4771 scope.go:117] "RemoveContainer" containerID="79a9cdd95ae64294ca9c39ad08d44cdc8883948e1e7f87ca95d709f164e260ab" Oct 02 11:58:12 crc kubenswrapper[4771]: I1002 11:58:12.690664 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 11:58:12 crc kubenswrapper[4771]: E1002 11:58:12.691105 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.433067 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wpjsh"] Oct 02 11:58:17 crc kubenswrapper[4771]: E1002 11:58:17.436558 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ead4b747-61c3-4cc5-8fc6-12313efb0ea4" containerName="extract-content" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.436587 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ead4b747-61c3-4cc5-8fc6-12313efb0ea4" containerName="extract-content" Oct 02 11:58:17 crc kubenswrapper[4771]: E1002 11:58:17.436625 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e395ad53-db6d-4495-9bde-13e6397d4c5b" containerName="registry-server" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.436633 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e395ad53-db6d-4495-9bde-13e6397d4c5b" containerName="registry-server" Oct 02 11:58:17 crc kubenswrapper[4771]: E1002 11:58:17.436663 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e395ad53-db6d-4495-9bde-13e6397d4c5b" containerName="extract-content" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.436669 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e395ad53-db6d-4495-9bde-13e6397d4c5b" containerName="extract-content" Oct 02 11:58:17 crc kubenswrapper[4771]: E1002 11:58:17.436683 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ead4b747-61c3-4cc5-8fc6-12313efb0ea4" containerName="extract-utilities" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.436690 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ead4b747-61c3-4cc5-8fc6-12313efb0ea4" containerName="extract-utilities" Oct 02 11:58:17 crc kubenswrapper[4771]: E1002 11:58:17.436705 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ead4b747-61c3-4cc5-8fc6-12313efb0ea4" containerName="registry-server" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.436711 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="ead4b747-61c3-4cc5-8fc6-12313efb0ea4" containerName="registry-server" Oct 02 11:58:17 crc kubenswrapper[4771]: E1002 11:58:17.436728 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e395ad53-db6d-4495-9bde-13e6397d4c5b" containerName="extract-utilities" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.436736 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="e395ad53-db6d-4495-9bde-13e6397d4c5b" containerName="extract-utilities" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.436979 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="e395ad53-db6d-4495-9bde-13e6397d4c5b" containerName="registry-server" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.437002 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="ead4b747-61c3-4cc5-8fc6-12313efb0ea4" containerName="registry-server" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.440949 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.458032 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wpjsh"] Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.577678 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-utilities\") pod \"redhat-operators-wpjsh\" (UID: \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\") " pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.578791 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-catalog-content\") pod \"redhat-operators-wpjsh\" (UID: \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\") " pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.579004 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8fxl\" (UniqueName: \"kubernetes.io/projected/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-kube-api-access-s8fxl\") pod \"redhat-operators-wpjsh\" (UID: \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\") " pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.682191 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-utilities\") pod \"redhat-operators-wpjsh\" (UID: \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\") " pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.682377 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-catalog-content\") pod \"redhat-operators-wpjsh\" (UID: \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\") " pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.682448 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8fxl\" (UniqueName: \"kubernetes.io/projected/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-kube-api-access-s8fxl\") pod \"redhat-operators-wpjsh\" (UID: \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\") " pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.684551 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-catalog-content\") pod \"redhat-operators-wpjsh\" (UID: \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\") " pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.684884 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-utilities\") pod \"redhat-operators-wpjsh\" (UID: \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\") " pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.728389 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8fxl\" (UniqueName: \"kubernetes.io/projected/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-kube-api-access-s8fxl\") pod \"redhat-operators-wpjsh\" (UID: \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\") " pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:17 crc kubenswrapper[4771]: I1002 11:58:17.773825 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:18 crc kubenswrapper[4771]: I1002 11:58:18.396535 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wpjsh"] Oct 02 11:58:18 crc kubenswrapper[4771]: I1002 11:58:18.799773 4771 generic.go:334] "Generic (PLEG): container finished" podID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerID="e2534d540f99c422e5c8dee3fbe304df416aea3fb6cdc97fa6f1dc0d33d0653c" exitCode=0 Oct 02 11:58:18 crc kubenswrapper[4771]: I1002 11:58:18.799847 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpjsh" event={"ID":"5a87aded-d7ea-41e5-9b28-c7baaf80aea4","Type":"ContainerDied","Data":"e2534d540f99c422e5c8dee3fbe304df416aea3fb6cdc97fa6f1dc0d33d0653c"} Oct 02 11:58:18 crc kubenswrapper[4771]: I1002 11:58:18.800137 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpjsh" event={"ID":"5a87aded-d7ea-41e5-9b28-c7baaf80aea4","Type":"ContainerStarted","Data":"f25fe55791055b46e8a15a5a64dd6dc94c5e35db442db20e07ed87f8142c66c6"} Oct 02 11:58:20 crc kubenswrapper[4771]: I1002 11:58:20.830323 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpjsh" event={"ID":"5a87aded-d7ea-41e5-9b28-c7baaf80aea4","Type":"ContainerStarted","Data":"50b775cc385d9f12e89bedd4b2639cd6a617ea89192fcfcf508900eafd850bd2"} Oct 02 11:58:25 crc kubenswrapper[4771]: I1002 11:58:25.919389 4771 generic.go:334] "Generic (PLEG): container finished" podID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerID="50b775cc385d9f12e89bedd4b2639cd6a617ea89192fcfcf508900eafd850bd2" exitCode=0 Oct 02 11:58:25 crc kubenswrapper[4771]: I1002 11:58:25.919597 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpjsh" event={"ID":"5a87aded-d7ea-41e5-9b28-c7baaf80aea4","Type":"ContainerDied","Data":"50b775cc385d9f12e89bedd4b2639cd6a617ea89192fcfcf508900eafd850bd2"} Oct 02 11:58:26 crc kubenswrapper[4771]: I1002 11:58:26.945270 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpjsh" event={"ID":"5a87aded-d7ea-41e5-9b28-c7baaf80aea4","Type":"ContainerStarted","Data":"1632e43e3dd22b56bfeff665a48f3fcf827704284e008a5a77b5899f95697de8"} Oct 02 11:58:26 crc kubenswrapper[4771]: I1002 11:58:26.984713 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wpjsh" podStartSLOduration=2.464726322 podStartE2EDuration="9.984688288s" podCreationTimestamp="2025-10-02 11:58:17 +0000 UTC" firstStartedPulling="2025-10-02 11:58:18.802511063 +0000 UTC m=+8486.450196130" lastFinishedPulling="2025-10-02 11:58:26.322473029 +0000 UTC m=+8493.970158096" observedRunningTime="2025-10-02 11:58:26.975979943 +0000 UTC m=+8494.623665010" watchObservedRunningTime="2025-10-02 11:58:26.984688288 +0000 UTC m=+8494.632373355" Oct 02 11:58:27 crc kubenswrapper[4771]: I1002 11:58:27.682605 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 11:58:27 crc kubenswrapper[4771]: E1002 11:58:27.683186 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:58:27 crc kubenswrapper[4771]: I1002 11:58:27.775328 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:27 crc kubenswrapper[4771]: I1002 11:58:27.775391 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:28 crc kubenswrapper[4771]: I1002 11:58:28.854681 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wpjsh" podUID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerName="registry-server" probeResult="failure" output=< Oct 02 11:58:28 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:58:28 crc kubenswrapper[4771]: > Oct 02 11:58:38 crc kubenswrapper[4771]: I1002 11:58:38.682763 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 11:58:38 crc kubenswrapper[4771]: E1002 11:58:38.684197 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:58:38 crc kubenswrapper[4771]: I1002 11:58:38.847875 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wpjsh" podUID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerName="registry-server" probeResult="failure" output=< Oct 02 11:58:38 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:58:38 crc kubenswrapper[4771]: > Oct 02 11:58:48 crc kubenswrapper[4771]: I1002 11:58:48.844816 4771 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wpjsh" podUID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerName="registry-server" probeResult="failure" output=< Oct 02 11:58:48 crc kubenswrapper[4771]: timeout: failed to connect service ":50051" within 1s Oct 02 11:58:48 crc kubenswrapper[4771]: > Oct 02 11:58:52 crc kubenswrapper[4771]: I1002 11:58:52.529979 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8xx6b"] Oct 02 11:58:52 crc kubenswrapper[4771]: I1002 11:58:52.536670 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:58:52 crc kubenswrapper[4771]: I1002 11:58:52.565764 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8xx6b"] Oct 02 11:58:52 crc kubenswrapper[4771]: I1002 11:58:52.675674 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/091a5789-22b6-43d2-9204-482a7a19a268-utilities\") pod \"community-operators-8xx6b\" (UID: \"091a5789-22b6-43d2-9204-482a7a19a268\") " pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:58:52 crc kubenswrapper[4771]: I1002 11:58:52.675896 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/091a5789-22b6-43d2-9204-482a7a19a268-catalog-content\") pod \"community-operators-8xx6b\" (UID: \"091a5789-22b6-43d2-9204-482a7a19a268\") " pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:58:52 crc kubenswrapper[4771]: I1002 11:58:52.676673 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blfjr\" (UniqueName: \"kubernetes.io/projected/091a5789-22b6-43d2-9204-482a7a19a268-kube-api-access-blfjr\") pod \"community-operators-8xx6b\" (UID: \"091a5789-22b6-43d2-9204-482a7a19a268\") " pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:58:52 crc kubenswrapper[4771]: I1002 11:58:52.779867 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/091a5789-22b6-43d2-9204-482a7a19a268-utilities\") pod \"community-operators-8xx6b\" (UID: \"091a5789-22b6-43d2-9204-482a7a19a268\") " pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:58:52 crc kubenswrapper[4771]: I1002 11:58:52.780035 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/091a5789-22b6-43d2-9204-482a7a19a268-catalog-content\") pod \"community-operators-8xx6b\" (UID: \"091a5789-22b6-43d2-9204-482a7a19a268\") " pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:58:52 crc kubenswrapper[4771]: I1002 11:58:52.780331 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blfjr\" (UniqueName: \"kubernetes.io/projected/091a5789-22b6-43d2-9204-482a7a19a268-kube-api-access-blfjr\") pod \"community-operators-8xx6b\" (UID: \"091a5789-22b6-43d2-9204-482a7a19a268\") " pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:58:52 crc kubenswrapper[4771]: I1002 11:58:52.782894 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/091a5789-22b6-43d2-9204-482a7a19a268-utilities\") pod \"community-operators-8xx6b\" (UID: \"091a5789-22b6-43d2-9204-482a7a19a268\") " pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:58:52 crc kubenswrapper[4771]: I1002 11:58:52.783071 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/091a5789-22b6-43d2-9204-482a7a19a268-catalog-content\") pod \"community-operators-8xx6b\" (UID: \"091a5789-22b6-43d2-9204-482a7a19a268\") " pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:58:52 crc kubenswrapper[4771]: I1002 11:58:52.805867 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blfjr\" (UniqueName: \"kubernetes.io/projected/091a5789-22b6-43d2-9204-482a7a19a268-kube-api-access-blfjr\") pod \"community-operators-8xx6b\" (UID: \"091a5789-22b6-43d2-9204-482a7a19a268\") " pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:58:52 crc kubenswrapper[4771]: I1002 11:58:52.862583 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:58:53 crc kubenswrapper[4771]: I1002 11:58:53.698694 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 11:58:53 crc kubenswrapper[4771]: E1002 11:58:53.700586 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:58:53 crc kubenswrapper[4771]: I1002 11:58:53.724809 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8xx6b"] Oct 02 11:58:54 crc kubenswrapper[4771]: I1002 11:58:54.453877 4771 generic.go:334] "Generic (PLEG): container finished" podID="091a5789-22b6-43d2-9204-482a7a19a268" containerID="52e78b414e83c13256e5ebc18fe6941b7742a368109fa2f85eb0ce8ae100dd2b" exitCode=0 Oct 02 11:58:54 crc kubenswrapper[4771]: I1002 11:58:54.453980 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xx6b" event={"ID":"091a5789-22b6-43d2-9204-482a7a19a268","Type":"ContainerDied","Data":"52e78b414e83c13256e5ebc18fe6941b7742a368109fa2f85eb0ce8ae100dd2b"} Oct 02 11:58:54 crc kubenswrapper[4771]: I1002 11:58:54.454407 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xx6b" event={"ID":"091a5789-22b6-43d2-9204-482a7a19a268","Type":"ContainerStarted","Data":"c57d717992c212a2f897c78202b5d99ca562135c5bc35a90f083ec516e3a0b89"} Oct 02 11:58:56 crc kubenswrapper[4771]: I1002 11:58:56.484986 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xx6b" event={"ID":"091a5789-22b6-43d2-9204-482a7a19a268","Type":"ContainerStarted","Data":"bdc002d718f55496f6a4704ab624587e0c4ca94805b3ba2a62f26875a6c691c0"} Oct 02 11:58:57 crc kubenswrapper[4771]: I1002 11:58:57.500480 4771 generic.go:334] "Generic (PLEG): container finished" podID="091a5789-22b6-43d2-9204-482a7a19a268" containerID="bdc002d718f55496f6a4704ab624587e0c4ca94805b3ba2a62f26875a6c691c0" exitCode=0 Oct 02 11:58:57 crc kubenswrapper[4771]: I1002 11:58:57.502580 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xx6b" event={"ID":"091a5789-22b6-43d2-9204-482a7a19a268","Type":"ContainerDied","Data":"bdc002d718f55496f6a4704ab624587e0c4ca94805b3ba2a62f26875a6c691c0"} Oct 02 11:58:57 crc kubenswrapper[4771]: I1002 11:58:57.845537 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:57 crc kubenswrapper[4771]: I1002 11:58:57.933775 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:58:58 crc kubenswrapper[4771]: I1002 11:58:58.516800 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xx6b" event={"ID":"091a5789-22b6-43d2-9204-482a7a19a268","Type":"ContainerStarted","Data":"9510f841331b32fa3617b578d344610d2050647691120294f33fd71b3c3600d8"} Oct 02 11:58:58 crc kubenswrapper[4771]: I1002 11:58:58.551244 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8xx6b" podStartSLOduration=3.094760052 podStartE2EDuration="6.551207718s" podCreationTimestamp="2025-10-02 11:58:52 +0000 UTC" firstStartedPulling="2025-10-02 11:58:54.456471172 +0000 UTC m=+8522.104156249" lastFinishedPulling="2025-10-02 11:58:57.912918848 +0000 UTC m=+8525.560603915" observedRunningTime="2025-10-02 11:58:58.542407151 +0000 UTC m=+8526.190092218" watchObservedRunningTime="2025-10-02 11:58:58.551207718 +0000 UTC m=+8526.198892795" Oct 02 11:58:59 crc kubenswrapper[4771]: I1002 11:58:59.104483 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wpjsh"] Oct 02 11:58:59 crc kubenswrapper[4771]: I1002 11:58:59.528450 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wpjsh" podUID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerName="registry-server" containerID="cri-o://1632e43e3dd22b56bfeff665a48f3fcf827704284e008a5a77b5899f95697de8" gracePeriod=2 Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.226250 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.353548 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-utilities\") pod \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\" (UID: \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\") " Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.354079 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8fxl\" (UniqueName: \"kubernetes.io/projected/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-kube-api-access-s8fxl\") pod \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\" (UID: \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\") " Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.354552 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-catalog-content\") pod \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\" (UID: \"5a87aded-d7ea-41e5-9b28-c7baaf80aea4\") " Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.354683 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-utilities" (OuterVolumeSpecName: "utilities") pod "5a87aded-d7ea-41e5-9b28-c7baaf80aea4" (UID: "5a87aded-d7ea-41e5-9b28-c7baaf80aea4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.355563 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.369696 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-kube-api-access-s8fxl" (OuterVolumeSpecName: "kube-api-access-s8fxl") pod "5a87aded-d7ea-41e5-9b28-c7baaf80aea4" (UID: "5a87aded-d7ea-41e5-9b28-c7baaf80aea4"). InnerVolumeSpecName "kube-api-access-s8fxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.459194 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8fxl\" (UniqueName: \"kubernetes.io/projected/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-kube-api-access-s8fxl\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.467550 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5a87aded-d7ea-41e5-9b28-c7baaf80aea4" (UID: "5a87aded-d7ea-41e5-9b28-c7baaf80aea4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.558387 4771 generic.go:334] "Generic (PLEG): container finished" podID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerID="1632e43e3dd22b56bfeff665a48f3fcf827704284e008a5a77b5899f95697de8" exitCode=0 Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.558464 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpjsh" event={"ID":"5a87aded-d7ea-41e5-9b28-c7baaf80aea4","Type":"ContainerDied","Data":"1632e43e3dd22b56bfeff665a48f3fcf827704284e008a5a77b5899f95697de8"} Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.558482 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wpjsh" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.558520 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpjsh" event={"ID":"5a87aded-d7ea-41e5-9b28-c7baaf80aea4","Type":"ContainerDied","Data":"f25fe55791055b46e8a15a5a64dd6dc94c5e35db442db20e07ed87f8142c66c6"} Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.558547 4771 scope.go:117] "RemoveContainer" containerID="1632e43e3dd22b56bfeff665a48f3fcf827704284e008a5a77b5899f95697de8" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.562208 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a87aded-d7ea-41e5-9b28-c7baaf80aea4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.595540 4771 scope.go:117] "RemoveContainer" containerID="50b775cc385d9f12e89bedd4b2639cd6a617ea89192fcfcf508900eafd850bd2" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.618357 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wpjsh"] Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.685718 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wpjsh"] Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.697506 4771 scope.go:117] "RemoveContainer" containerID="e2534d540f99c422e5c8dee3fbe304df416aea3fb6cdc97fa6f1dc0d33d0653c" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.754623 4771 scope.go:117] "RemoveContainer" containerID="1632e43e3dd22b56bfeff665a48f3fcf827704284e008a5a77b5899f95697de8" Oct 02 11:59:00 crc kubenswrapper[4771]: E1002 11:59:00.755415 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1632e43e3dd22b56bfeff665a48f3fcf827704284e008a5a77b5899f95697de8\": container with ID starting with 1632e43e3dd22b56bfeff665a48f3fcf827704284e008a5a77b5899f95697de8 not found: ID does not exist" containerID="1632e43e3dd22b56bfeff665a48f3fcf827704284e008a5a77b5899f95697de8" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.755491 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1632e43e3dd22b56bfeff665a48f3fcf827704284e008a5a77b5899f95697de8"} err="failed to get container status \"1632e43e3dd22b56bfeff665a48f3fcf827704284e008a5a77b5899f95697de8\": rpc error: code = NotFound desc = could not find container \"1632e43e3dd22b56bfeff665a48f3fcf827704284e008a5a77b5899f95697de8\": container with ID starting with 1632e43e3dd22b56bfeff665a48f3fcf827704284e008a5a77b5899f95697de8 not found: ID does not exist" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.755535 4771 scope.go:117] "RemoveContainer" containerID="50b775cc385d9f12e89bedd4b2639cd6a617ea89192fcfcf508900eafd850bd2" Oct 02 11:59:00 crc kubenswrapper[4771]: E1002 11:59:00.755906 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50b775cc385d9f12e89bedd4b2639cd6a617ea89192fcfcf508900eafd850bd2\": container with ID starting with 50b775cc385d9f12e89bedd4b2639cd6a617ea89192fcfcf508900eafd850bd2 not found: ID does not exist" containerID="50b775cc385d9f12e89bedd4b2639cd6a617ea89192fcfcf508900eafd850bd2" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.755934 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50b775cc385d9f12e89bedd4b2639cd6a617ea89192fcfcf508900eafd850bd2"} err="failed to get container status \"50b775cc385d9f12e89bedd4b2639cd6a617ea89192fcfcf508900eafd850bd2\": rpc error: code = NotFound desc = could not find container \"50b775cc385d9f12e89bedd4b2639cd6a617ea89192fcfcf508900eafd850bd2\": container with ID starting with 50b775cc385d9f12e89bedd4b2639cd6a617ea89192fcfcf508900eafd850bd2 not found: ID does not exist" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.755953 4771 scope.go:117] "RemoveContainer" containerID="e2534d540f99c422e5c8dee3fbe304df416aea3fb6cdc97fa6f1dc0d33d0653c" Oct 02 11:59:00 crc kubenswrapper[4771]: E1002 11:59:00.756246 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2534d540f99c422e5c8dee3fbe304df416aea3fb6cdc97fa6f1dc0d33d0653c\": container with ID starting with e2534d540f99c422e5c8dee3fbe304df416aea3fb6cdc97fa6f1dc0d33d0653c not found: ID does not exist" containerID="e2534d540f99c422e5c8dee3fbe304df416aea3fb6cdc97fa6f1dc0d33d0653c" Oct 02 11:59:00 crc kubenswrapper[4771]: I1002 11:59:00.756280 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2534d540f99c422e5c8dee3fbe304df416aea3fb6cdc97fa6f1dc0d33d0653c"} err="failed to get container status \"e2534d540f99c422e5c8dee3fbe304df416aea3fb6cdc97fa6f1dc0d33d0653c\": rpc error: code = NotFound desc = could not find container \"e2534d540f99c422e5c8dee3fbe304df416aea3fb6cdc97fa6f1dc0d33d0653c\": container with ID starting with e2534d540f99c422e5c8dee3fbe304df416aea3fb6cdc97fa6f1dc0d33d0653c not found: ID does not exist" Oct 02 11:59:01 crc kubenswrapper[4771]: I1002 11:59:01.703483 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" path="/var/lib/kubelet/pods/5a87aded-d7ea-41e5-9b28-c7baaf80aea4/volumes" Oct 02 11:59:02 crc kubenswrapper[4771]: I1002 11:59:02.863999 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:59:02 crc kubenswrapper[4771]: I1002 11:59:02.864090 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:59:02 crc kubenswrapper[4771]: I1002 11:59:02.943410 4771 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:59:03 crc kubenswrapper[4771]: I1002 11:59:03.877676 4771 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:59:04 crc kubenswrapper[4771]: I1002 11:59:04.099840 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8xx6b"] Oct 02 11:59:04 crc kubenswrapper[4771]: I1002 11:59:04.682006 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 11:59:04 crc kubenswrapper[4771]: E1002 11:59:04.682685 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:59:05 crc kubenswrapper[4771]: I1002 11:59:05.720846 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8xx6b" podUID="091a5789-22b6-43d2-9204-482a7a19a268" containerName="registry-server" containerID="cri-o://9510f841331b32fa3617b578d344610d2050647691120294f33fd71b3c3600d8" gracePeriod=2 Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.405090 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.527211 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/091a5789-22b6-43d2-9204-482a7a19a268-catalog-content\") pod \"091a5789-22b6-43d2-9204-482a7a19a268\" (UID: \"091a5789-22b6-43d2-9204-482a7a19a268\") " Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.527440 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blfjr\" (UniqueName: \"kubernetes.io/projected/091a5789-22b6-43d2-9204-482a7a19a268-kube-api-access-blfjr\") pod \"091a5789-22b6-43d2-9204-482a7a19a268\" (UID: \"091a5789-22b6-43d2-9204-482a7a19a268\") " Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.527622 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/091a5789-22b6-43d2-9204-482a7a19a268-utilities\") pod \"091a5789-22b6-43d2-9204-482a7a19a268\" (UID: \"091a5789-22b6-43d2-9204-482a7a19a268\") " Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.528648 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/091a5789-22b6-43d2-9204-482a7a19a268-utilities" (OuterVolumeSpecName: "utilities") pod "091a5789-22b6-43d2-9204-482a7a19a268" (UID: "091a5789-22b6-43d2-9204-482a7a19a268"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.529631 4771 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/091a5789-22b6-43d2-9204-482a7a19a268-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.542432 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/091a5789-22b6-43d2-9204-482a7a19a268-kube-api-access-blfjr" (OuterVolumeSpecName: "kube-api-access-blfjr") pod "091a5789-22b6-43d2-9204-482a7a19a268" (UID: "091a5789-22b6-43d2-9204-482a7a19a268"). InnerVolumeSpecName "kube-api-access-blfjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.588678 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/091a5789-22b6-43d2-9204-482a7a19a268-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "091a5789-22b6-43d2-9204-482a7a19a268" (UID: "091a5789-22b6-43d2-9204-482a7a19a268"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.633212 4771 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/091a5789-22b6-43d2-9204-482a7a19a268-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.633270 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blfjr\" (UniqueName: \"kubernetes.io/projected/091a5789-22b6-43d2-9204-482a7a19a268-kube-api-access-blfjr\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.737510 4771 generic.go:334] "Generic (PLEG): container finished" podID="091a5789-22b6-43d2-9204-482a7a19a268" containerID="9510f841331b32fa3617b578d344610d2050647691120294f33fd71b3c3600d8" exitCode=0 Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.737575 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xx6b" event={"ID":"091a5789-22b6-43d2-9204-482a7a19a268","Type":"ContainerDied","Data":"9510f841331b32fa3617b578d344610d2050647691120294f33fd71b3c3600d8"} Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.737614 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xx6b" event={"ID":"091a5789-22b6-43d2-9204-482a7a19a268","Type":"ContainerDied","Data":"c57d717992c212a2f897c78202b5d99ca562135c5bc35a90f083ec516e3a0b89"} Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.737637 4771 scope.go:117] "RemoveContainer" containerID="9510f841331b32fa3617b578d344610d2050647691120294f33fd71b3c3600d8" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.737840 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xx6b" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.782002 4771 scope.go:117] "RemoveContainer" containerID="bdc002d718f55496f6a4704ab624587e0c4ca94805b3ba2a62f26875a6c691c0" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.782733 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8xx6b"] Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.795572 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8xx6b"] Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.833962 4771 scope.go:117] "RemoveContainer" containerID="52e78b414e83c13256e5ebc18fe6941b7742a368109fa2f85eb0ce8ae100dd2b" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.881340 4771 scope.go:117] "RemoveContainer" containerID="9510f841331b32fa3617b578d344610d2050647691120294f33fd71b3c3600d8" Oct 02 11:59:06 crc kubenswrapper[4771]: E1002 11:59:06.888622 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9510f841331b32fa3617b578d344610d2050647691120294f33fd71b3c3600d8\": container with ID starting with 9510f841331b32fa3617b578d344610d2050647691120294f33fd71b3c3600d8 not found: ID does not exist" containerID="9510f841331b32fa3617b578d344610d2050647691120294f33fd71b3c3600d8" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.888712 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9510f841331b32fa3617b578d344610d2050647691120294f33fd71b3c3600d8"} err="failed to get container status \"9510f841331b32fa3617b578d344610d2050647691120294f33fd71b3c3600d8\": rpc error: code = NotFound desc = could not find container \"9510f841331b32fa3617b578d344610d2050647691120294f33fd71b3c3600d8\": container with ID starting with 9510f841331b32fa3617b578d344610d2050647691120294f33fd71b3c3600d8 not found: ID does not exist" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.888751 4771 scope.go:117] "RemoveContainer" containerID="bdc002d718f55496f6a4704ab624587e0c4ca94805b3ba2a62f26875a6c691c0" Oct 02 11:59:06 crc kubenswrapper[4771]: E1002 11:59:06.889639 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdc002d718f55496f6a4704ab624587e0c4ca94805b3ba2a62f26875a6c691c0\": container with ID starting with bdc002d718f55496f6a4704ab624587e0c4ca94805b3ba2a62f26875a6c691c0 not found: ID does not exist" containerID="bdc002d718f55496f6a4704ab624587e0c4ca94805b3ba2a62f26875a6c691c0" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.889682 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdc002d718f55496f6a4704ab624587e0c4ca94805b3ba2a62f26875a6c691c0"} err="failed to get container status \"bdc002d718f55496f6a4704ab624587e0c4ca94805b3ba2a62f26875a6c691c0\": rpc error: code = NotFound desc = could not find container \"bdc002d718f55496f6a4704ab624587e0c4ca94805b3ba2a62f26875a6c691c0\": container with ID starting with bdc002d718f55496f6a4704ab624587e0c4ca94805b3ba2a62f26875a6c691c0 not found: ID does not exist" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.889709 4771 scope.go:117] "RemoveContainer" containerID="52e78b414e83c13256e5ebc18fe6941b7742a368109fa2f85eb0ce8ae100dd2b" Oct 02 11:59:06 crc kubenswrapper[4771]: E1002 11:59:06.890380 4771 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52e78b414e83c13256e5ebc18fe6941b7742a368109fa2f85eb0ce8ae100dd2b\": container with ID starting with 52e78b414e83c13256e5ebc18fe6941b7742a368109fa2f85eb0ce8ae100dd2b not found: ID does not exist" containerID="52e78b414e83c13256e5ebc18fe6941b7742a368109fa2f85eb0ce8ae100dd2b" Oct 02 11:59:06 crc kubenswrapper[4771]: I1002 11:59:06.890495 4771 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52e78b414e83c13256e5ebc18fe6941b7742a368109fa2f85eb0ce8ae100dd2b"} err="failed to get container status \"52e78b414e83c13256e5ebc18fe6941b7742a368109fa2f85eb0ce8ae100dd2b\": rpc error: code = NotFound desc = could not find container \"52e78b414e83c13256e5ebc18fe6941b7742a368109fa2f85eb0ce8ae100dd2b\": container with ID starting with 52e78b414e83c13256e5ebc18fe6941b7742a368109fa2f85eb0ce8ae100dd2b not found: ID does not exist" Oct 02 11:59:07 crc kubenswrapper[4771]: I1002 11:59:07.699748 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="091a5789-22b6-43d2-9204-482a7a19a268" path="/var/lib/kubelet/pods/091a5789-22b6-43d2-9204-482a7a19a268/volumes" Oct 02 11:59:19 crc kubenswrapper[4771]: I1002 11:59:19.684570 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 11:59:19 crc kubenswrapper[4771]: E1002 11:59:19.685463 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:59:32 crc kubenswrapper[4771]: I1002 11:59:32.681994 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 11:59:32 crc kubenswrapper[4771]: E1002 11:59:32.683229 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 11:59:47 crc kubenswrapper[4771]: I1002 11:59:47.681252 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 11:59:47 crc kubenswrapper[4771]: E1002 11:59:47.682303 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.250574 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb"] Oct 02 12:00:00 crc kubenswrapper[4771]: E1002 12:00:00.253382 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091a5789-22b6-43d2-9204-482a7a19a268" containerName="extract-content" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.253422 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="091a5789-22b6-43d2-9204-482a7a19a268" containerName="extract-content" Oct 02 12:00:00 crc kubenswrapper[4771]: E1002 12:00:00.253529 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerName="extract-content" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.253540 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerName="extract-content" Oct 02 12:00:00 crc kubenswrapper[4771]: E1002 12:00:00.253594 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091a5789-22b6-43d2-9204-482a7a19a268" containerName="extract-utilities" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.253610 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="091a5789-22b6-43d2-9204-482a7a19a268" containerName="extract-utilities" Oct 02 12:00:00 crc kubenswrapper[4771]: E1002 12:00:00.253638 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerName="registry-server" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.253648 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerName="registry-server" Oct 02 12:00:00 crc kubenswrapper[4771]: E1002 12:00:00.253701 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerName="extract-utilities" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.253711 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerName="extract-utilities" Oct 02 12:00:00 crc kubenswrapper[4771]: E1002 12:00:00.253742 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091a5789-22b6-43d2-9204-482a7a19a268" containerName="registry-server" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.253754 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="091a5789-22b6-43d2-9204-482a7a19a268" containerName="registry-server" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.254605 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a87aded-d7ea-41e5-9b28-c7baaf80aea4" containerName="registry-server" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.254653 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="091a5789-22b6-43d2-9204-482a7a19a268" containerName="registry-server" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.261718 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.290882 4771 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.291441 4771 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.305312 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb"] Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.417901 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0a2f3fd-5b84-4376-bf91-eea588004739-secret-volume\") pod \"collect-profiles-29323440-7mjtb\" (UID: \"a0a2f3fd-5b84-4376-bf91-eea588004739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.417975 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsb2g\" (UniqueName: \"kubernetes.io/projected/a0a2f3fd-5b84-4376-bf91-eea588004739-kube-api-access-jsb2g\") pod \"collect-profiles-29323440-7mjtb\" (UID: \"a0a2f3fd-5b84-4376-bf91-eea588004739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.418030 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0a2f3fd-5b84-4376-bf91-eea588004739-config-volume\") pod \"collect-profiles-29323440-7mjtb\" (UID: \"a0a2f3fd-5b84-4376-bf91-eea588004739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.521317 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0a2f3fd-5b84-4376-bf91-eea588004739-secret-volume\") pod \"collect-profiles-29323440-7mjtb\" (UID: \"a0a2f3fd-5b84-4376-bf91-eea588004739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.521634 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsb2g\" (UniqueName: \"kubernetes.io/projected/a0a2f3fd-5b84-4376-bf91-eea588004739-kube-api-access-jsb2g\") pod \"collect-profiles-29323440-7mjtb\" (UID: \"a0a2f3fd-5b84-4376-bf91-eea588004739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.521680 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0a2f3fd-5b84-4376-bf91-eea588004739-config-volume\") pod \"collect-profiles-29323440-7mjtb\" (UID: \"a0a2f3fd-5b84-4376-bf91-eea588004739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.522948 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0a2f3fd-5b84-4376-bf91-eea588004739-config-volume\") pod \"collect-profiles-29323440-7mjtb\" (UID: \"a0a2f3fd-5b84-4376-bf91-eea588004739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.531847 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0a2f3fd-5b84-4376-bf91-eea588004739-secret-volume\") pod \"collect-profiles-29323440-7mjtb\" (UID: \"a0a2f3fd-5b84-4376-bf91-eea588004739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.549509 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsb2g\" (UniqueName: \"kubernetes.io/projected/a0a2f3fd-5b84-4376-bf91-eea588004739-kube-api-access-jsb2g\") pod \"collect-profiles-29323440-7mjtb\" (UID: \"a0a2f3fd-5b84-4376-bf91-eea588004739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.605780 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" Oct 02 12:00:00 crc kubenswrapper[4771]: I1002 12:00:00.682126 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:00:00 crc kubenswrapper[4771]: E1002 12:00:00.682610 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:00:01 crc kubenswrapper[4771]: I1002 12:00:01.175471 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb"] Oct 02 12:00:01 crc kubenswrapper[4771]: I1002 12:00:01.555442 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" event={"ID":"a0a2f3fd-5b84-4376-bf91-eea588004739","Type":"ContainerStarted","Data":"38b08f0970d046e9809c6b8d7d2ae1e7031e24f021d8d17bb08e82236a88fef5"} Oct 02 12:00:01 crc kubenswrapper[4771]: I1002 12:00:01.555883 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" event={"ID":"a0a2f3fd-5b84-4376-bf91-eea588004739","Type":"ContainerStarted","Data":"e0b6bc3c1cffd9b00731fa692c304c563edd2be2344745157fa3f8e1121e10cc"} Oct 02 12:00:01 crc kubenswrapper[4771]: I1002 12:00:01.584097 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" podStartSLOduration=1.584063392 podStartE2EDuration="1.584063392s" podCreationTimestamp="2025-10-02 12:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:00:01.577432511 +0000 UTC m=+8589.225117588" watchObservedRunningTime="2025-10-02 12:00:01.584063392 +0000 UTC m=+8589.231748499" Oct 02 12:00:02 crc kubenswrapper[4771]: I1002 12:00:02.576290 4771 generic.go:334] "Generic (PLEG): container finished" podID="a0a2f3fd-5b84-4376-bf91-eea588004739" containerID="38b08f0970d046e9809c6b8d7d2ae1e7031e24f021d8d17bb08e82236a88fef5" exitCode=0 Oct 02 12:00:02 crc kubenswrapper[4771]: I1002 12:00:02.576458 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" event={"ID":"a0a2f3fd-5b84-4376-bf91-eea588004739","Type":"ContainerDied","Data":"38b08f0970d046e9809c6b8d7d2ae1e7031e24f021d8d17bb08e82236a88fef5"} Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.128428 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.266099 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0a2f3fd-5b84-4376-bf91-eea588004739-config-volume\") pod \"a0a2f3fd-5b84-4376-bf91-eea588004739\" (UID: \"a0a2f3fd-5b84-4376-bf91-eea588004739\") " Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.266503 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsb2g\" (UniqueName: \"kubernetes.io/projected/a0a2f3fd-5b84-4376-bf91-eea588004739-kube-api-access-jsb2g\") pod \"a0a2f3fd-5b84-4376-bf91-eea588004739\" (UID: \"a0a2f3fd-5b84-4376-bf91-eea588004739\") " Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.266708 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0a2f3fd-5b84-4376-bf91-eea588004739-secret-volume\") pod \"a0a2f3fd-5b84-4376-bf91-eea588004739\" (UID: \"a0a2f3fd-5b84-4376-bf91-eea588004739\") " Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.267672 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0a2f3fd-5b84-4376-bf91-eea588004739-config-volume" (OuterVolumeSpecName: "config-volume") pod "a0a2f3fd-5b84-4376-bf91-eea588004739" (UID: "a0a2f3fd-5b84-4376-bf91-eea588004739"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.274287 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0a2f3fd-5b84-4376-bf91-eea588004739-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a0a2f3fd-5b84-4376-bf91-eea588004739" (UID: "a0a2f3fd-5b84-4376-bf91-eea588004739"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.296699 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0a2f3fd-5b84-4376-bf91-eea588004739-kube-api-access-jsb2g" (OuterVolumeSpecName: "kube-api-access-jsb2g") pod "a0a2f3fd-5b84-4376-bf91-eea588004739" (UID: "a0a2f3fd-5b84-4376-bf91-eea588004739"). InnerVolumeSpecName "kube-api-access-jsb2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.370668 4771 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0a2f3fd-5b84-4376-bf91-eea588004739-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.370715 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsb2g\" (UniqueName: \"kubernetes.io/projected/a0a2f3fd-5b84-4376-bf91-eea588004739-kube-api-access-jsb2g\") on node \"crc\" DevicePath \"\"" Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.370726 4771 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0a2f3fd-5b84-4376-bf91-eea588004739-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.631546 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" event={"ID":"a0a2f3fd-5b84-4376-bf91-eea588004739","Type":"ContainerDied","Data":"e0b6bc3c1cffd9b00731fa692c304c563edd2be2344745157fa3f8e1121e10cc"} Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.631635 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0b6bc3c1cffd9b00731fa692c304c563edd2be2344745157fa3f8e1121e10cc" Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.631630 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-7mjtb" Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.706082 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw"] Oct 02 12:00:04 crc kubenswrapper[4771]: I1002 12:00:04.719449 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323395-xmfpw"] Oct 02 12:00:05 crc kubenswrapper[4771]: I1002 12:00:05.710757 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36b7c50d-0488-4ad1-b2e5-756416f88ca6" path="/var/lib/kubelet/pods/36b7c50d-0488-4ad1-b2e5-756416f88ca6/volumes" Oct 02 12:00:09 crc kubenswrapper[4771]: I1002 12:00:09.397178 4771 scope.go:117] "RemoveContainer" containerID="112057a734f698c94d1172b43104c19bfa1bbfa09851d4cc26ca3dda69b2b7e2" Oct 02 12:00:14 crc kubenswrapper[4771]: I1002 12:00:14.681639 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:00:14 crc kubenswrapper[4771]: E1002 12:00:14.682624 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:00:26 crc kubenswrapper[4771]: I1002 12:00:26.686996 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:00:26 crc kubenswrapper[4771]: E1002 12:00:26.688488 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:00:35 crc kubenswrapper[4771]: I1002 12:00:35.233077 4771 generic.go:334] "Generic (PLEG): container finished" podID="1e26af54-9be2-44da-9bbd-0684c864ce9e" containerID="21e072d36ab3899659bc92f035fc77bf6c5e73d35b8d61fa026f411d0baf4ee0" exitCode=0 Oct 02 12:00:35 crc kubenswrapper[4771]: I1002 12:00:35.233804 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pc9ng/must-gather-j92sz" event={"ID":"1e26af54-9be2-44da-9bbd-0684c864ce9e","Type":"ContainerDied","Data":"21e072d36ab3899659bc92f035fc77bf6c5e73d35b8d61fa026f411d0baf4ee0"} Oct 02 12:00:35 crc kubenswrapper[4771]: I1002 12:00:35.235097 4771 scope.go:117] "RemoveContainer" containerID="21e072d36ab3899659bc92f035fc77bf6c5e73d35b8d61fa026f411d0baf4ee0" Oct 02 12:00:35 crc kubenswrapper[4771]: I1002 12:00:35.530433 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pc9ng_must-gather-j92sz_1e26af54-9be2-44da-9bbd-0684c864ce9e/gather/0.log" Oct 02 12:00:38 crc kubenswrapper[4771]: I1002 12:00:38.688519 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:00:38 crc kubenswrapper[4771]: E1002 12:00:38.689678 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:00:48 crc kubenswrapper[4771]: I1002 12:00:48.633005 4771 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pc9ng/must-gather-j92sz"] Oct 02 12:00:48 crc kubenswrapper[4771]: I1002 12:00:48.634606 4771 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-pc9ng/must-gather-j92sz" podUID="1e26af54-9be2-44da-9bbd-0684c864ce9e" containerName="copy" containerID="cri-o://30e6fe3afab65597caa31dc9566ca49b39d65338a78357367c6cb3c4003713ee" gracePeriod=2 Oct 02 12:00:48 crc kubenswrapper[4771]: I1002 12:00:48.651621 4771 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pc9ng/must-gather-j92sz"] Oct 02 12:00:49 crc kubenswrapper[4771]: I1002 12:00:49.520430 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pc9ng_must-gather-j92sz_1e26af54-9be2-44da-9bbd-0684c864ce9e/copy/0.log" Oct 02 12:00:49 crc kubenswrapper[4771]: I1002 12:00:49.535902 4771 generic.go:334] "Generic (PLEG): container finished" podID="1e26af54-9be2-44da-9bbd-0684c864ce9e" containerID="30e6fe3afab65597caa31dc9566ca49b39d65338a78357367c6cb3c4003713ee" exitCode=143 Oct 02 12:00:49 crc kubenswrapper[4771]: I1002 12:00:49.535981 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad92f191061cdd447aa48134eede19edf0b2a667fd5ecacc6bd71e71f34c0a51" Oct 02 12:00:49 crc kubenswrapper[4771]: I1002 12:00:49.564044 4771 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pc9ng_must-gather-j92sz_1e26af54-9be2-44da-9bbd-0684c864ce9e/copy/0.log" Oct 02 12:00:49 crc kubenswrapper[4771]: I1002 12:00:49.571713 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/must-gather-j92sz" Oct 02 12:00:49 crc kubenswrapper[4771]: I1002 12:00:49.688203 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws8fg\" (UniqueName: \"kubernetes.io/projected/1e26af54-9be2-44da-9bbd-0684c864ce9e-kube-api-access-ws8fg\") pod \"1e26af54-9be2-44da-9bbd-0684c864ce9e\" (UID: \"1e26af54-9be2-44da-9bbd-0684c864ce9e\") " Oct 02 12:00:49 crc kubenswrapper[4771]: I1002 12:00:49.688953 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e26af54-9be2-44da-9bbd-0684c864ce9e-must-gather-output\") pod \"1e26af54-9be2-44da-9bbd-0684c864ce9e\" (UID: \"1e26af54-9be2-44da-9bbd-0684c864ce9e\") " Oct 02 12:00:49 crc kubenswrapper[4771]: I1002 12:00:49.760608 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e26af54-9be2-44da-9bbd-0684c864ce9e-kube-api-access-ws8fg" (OuterVolumeSpecName: "kube-api-access-ws8fg") pod "1e26af54-9be2-44da-9bbd-0684c864ce9e" (UID: "1e26af54-9be2-44da-9bbd-0684c864ce9e"). InnerVolumeSpecName "kube-api-access-ws8fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:00:49 crc kubenswrapper[4771]: I1002 12:00:49.801424 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws8fg\" (UniqueName: \"kubernetes.io/projected/1e26af54-9be2-44da-9bbd-0684c864ce9e-kube-api-access-ws8fg\") on node \"crc\" DevicePath \"\"" Oct 02 12:00:49 crc kubenswrapper[4771]: I1002 12:00:49.928469 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e26af54-9be2-44da-9bbd-0684c864ce9e-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "1e26af54-9be2-44da-9bbd-0684c864ce9e" (UID: "1e26af54-9be2-44da-9bbd-0684c864ce9e"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:00:50 crc kubenswrapper[4771]: I1002 12:00:50.007469 4771 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e26af54-9be2-44da-9bbd-0684c864ce9e-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 02 12:00:50 crc kubenswrapper[4771]: I1002 12:00:50.549191 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pc9ng/must-gather-j92sz" Oct 02 12:00:51 crc kubenswrapper[4771]: I1002 12:00:51.712292 4771 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e26af54-9be2-44da-9bbd-0684c864ce9e" path="/var/lib/kubelet/pods/1e26af54-9be2-44da-9bbd-0684c864ce9e/volumes" Oct 02 12:00:53 crc kubenswrapper[4771]: I1002 12:00:53.698112 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:00:53 crc kubenswrapper[4771]: E1002 12:00:53.701249 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.172967 4771 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29323441-5fw9w"] Oct 02 12:01:00 crc kubenswrapper[4771]: E1002 12:01:00.175348 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e26af54-9be2-44da-9bbd-0684c864ce9e" containerName="copy" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.175389 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e26af54-9be2-44da-9bbd-0684c864ce9e" containerName="copy" Oct 02 12:01:00 crc kubenswrapper[4771]: E1002 12:01:00.175426 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e26af54-9be2-44da-9bbd-0684c864ce9e" containerName="gather" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.175435 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e26af54-9be2-44da-9bbd-0684c864ce9e" containerName="gather" Oct 02 12:01:00 crc kubenswrapper[4771]: E1002 12:01:00.175459 4771 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0a2f3fd-5b84-4376-bf91-eea588004739" containerName="collect-profiles" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.175468 4771 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0a2f3fd-5b84-4376-bf91-eea588004739" containerName="collect-profiles" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.175869 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0a2f3fd-5b84-4376-bf91-eea588004739" containerName="collect-profiles" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.175898 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e26af54-9be2-44da-9bbd-0684c864ce9e" containerName="gather" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.175927 4771 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e26af54-9be2-44da-9bbd-0684c864ce9e" containerName="copy" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.177486 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.205395 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323441-5fw9w"] Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.336795 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-config-data\") pod \"keystone-cron-29323441-5fw9w\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.337315 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vntgq\" (UniqueName: \"kubernetes.io/projected/8333e13f-306d-4f64-aa36-78bef844a302-kube-api-access-vntgq\") pod \"keystone-cron-29323441-5fw9w\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.337511 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-combined-ca-bundle\") pod \"keystone-cron-29323441-5fw9w\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.337591 4771 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-fernet-keys\") pod \"keystone-cron-29323441-5fw9w\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.441988 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-config-data\") pod \"keystone-cron-29323441-5fw9w\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.442403 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vntgq\" (UniqueName: \"kubernetes.io/projected/8333e13f-306d-4f64-aa36-78bef844a302-kube-api-access-vntgq\") pod \"keystone-cron-29323441-5fw9w\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.442469 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-combined-ca-bundle\") pod \"keystone-cron-29323441-5fw9w\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.442505 4771 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-fernet-keys\") pod \"keystone-cron-29323441-5fw9w\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.454453 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-combined-ca-bundle\") pod \"keystone-cron-29323441-5fw9w\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.454928 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-config-data\") pod \"keystone-cron-29323441-5fw9w\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.455839 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-fernet-keys\") pod \"keystone-cron-29323441-5fw9w\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.465313 4771 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vntgq\" (UniqueName: \"kubernetes.io/projected/8333e13f-306d-4f64-aa36-78bef844a302-kube-api-access-vntgq\") pod \"keystone-cron-29323441-5fw9w\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:00 crc kubenswrapper[4771]: I1002 12:01:00.505756 4771 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:01 crc kubenswrapper[4771]: I1002 12:01:01.163350 4771 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323441-5fw9w"] Oct 02 12:01:01 crc kubenswrapper[4771]: I1002 12:01:01.769214 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323441-5fw9w" event={"ID":"8333e13f-306d-4f64-aa36-78bef844a302","Type":"ContainerStarted","Data":"2a0aa0edb6f7c55c8307f54bc471b6f2fecf6391e1da5f7447a6573c2acdc95c"} Oct 02 12:01:01 crc kubenswrapper[4771]: I1002 12:01:01.769637 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323441-5fw9w" event={"ID":"8333e13f-306d-4f64-aa36-78bef844a302","Type":"ContainerStarted","Data":"4bd8b4c6d3ba7273ba1c97715057f41c1fc8d02d126b9d2afdfa43032f176710"} Oct 02 12:01:01 crc kubenswrapper[4771]: I1002 12:01:01.797198 4771 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29323441-5fw9w" podStartSLOduration=1.7971712800000001 podStartE2EDuration="1.79717128s" podCreationTimestamp="2025-10-02 12:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:01:01.788016404 +0000 UTC m=+8649.435701471" watchObservedRunningTime="2025-10-02 12:01:01.79717128 +0000 UTC m=+8649.444856347" Oct 02 12:01:05 crc kubenswrapper[4771]: I1002 12:01:05.861550 4771 generic.go:334] "Generic (PLEG): container finished" podID="8333e13f-306d-4f64-aa36-78bef844a302" containerID="2a0aa0edb6f7c55c8307f54bc471b6f2fecf6391e1da5f7447a6573c2acdc95c" exitCode=0 Oct 02 12:01:05 crc kubenswrapper[4771]: I1002 12:01:05.861638 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323441-5fw9w" event={"ID":"8333e13f-306d-4f64-aa36-78bef844a302","Type":"ContainerDied","Data":"2a0aa0edb6f7c55c8307f54bc471b6f2fecf6391e1da5f7447a6573c2acdc95c"} Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.516542 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.672189 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vntgq\" (UniqueName: \"kubernetes.io/projected/8333e13f-306d-4f64-aa36-78bef844a302-kube-api-access-vntgq\") pod \"8333e13f-306d-4f64-aa36-78bef844a302\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.672288 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-config-data\") pod \"8333e13f-306d-4f64-aa36-78bef844a302\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.672356 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-combined-ca-bundle\") pod \"8333e13f-306d-4f64-aa36-78bef844a302\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.672586 4771 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-fernet-keys\") pod \"8333e13f-306d-4f64-aa36-78bef844a302\" (UID: \"8333e13f-306d-4f64-aa36-78bef844a302\") " Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.684224 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8333e13f-306d-4f64-aa36-78bef844a302-kube-api-access-vntgq" (OuterVolumeSpecName: "kube-api-access-vntgq") pod "8333e13f-306d-4f64-aa36-78bef844a302" (UID: "8333e13f-306d-4f64-aa36-78bef844a302"). InnerVolumeSpecName "kube-api-access-vntgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.692629 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8333e13f-306d-4f64-aa36-78bef844a302" (UID: "8333e13f-306d-4f64-aa36-78bef844a302"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.723039 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8333e13f-306d-4f64-aa36-78bef844a302" (UID: "8333e13f-306d-4f64-aa36-78bef844a302"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.784025 4771 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-config-data" (OuterVolumeSpecName: "config-data") pod "8333e13f-306d-4f64-aa36-78bef844a302" (UID: "8333e13f-306d-4f64-aa36-78bef844a302"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.785989 4771 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.786049 4771 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vntgq\" (UniqueName: \"kubernetes.io/projected/8333e13f-306d-4f64-aa36-78bef844a302-kube-api-access-vntgq\") on node \"crc\" DevicePath \"\"" Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.786073 4771 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.786092 4771 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8333e13f-306d-4f64-aa36-78bef844a302-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.897381 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323441-5fw9w" event={"ID":"8333e13f-306d-4f64-aa36-78bef844a302","Type":"ContainerDied","Data":"4bd8b4c6d3ba7273ba1c97715057f41c1fc8d02d126b9d2afdfa43032f176710"} Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.897440 4771 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bd8b4c6d3ba7273ba1c97715057f41c1fc8d02d126b9d2afdfa43032f176710" Oct 02 12:01:07 crc kubenswrapper[4771]: I1002 12:01:07.897541 4771 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323441-5fw9w" Oct 02 12:01:08 crc kubenswrapper[4771]: I1002 12:01:08.683837 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:01:08 crc kubenswrapper[4771]: E1002 12:01:08.685443 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:01:09 crc kubenswrapper[4771]: I1002 12:01:09.503305 4771 scope.go:117] "RemoveContainer" containerID="21e072d36ab3899659bc92f035fc77bf6c5e73d35b8d61fa026f411d0baf4ee0" Oct 02 12:01:09 crc kubenswrapper[4771]: I1002 12:01:09.659335 4771 scope.go:117] "RemoveContainer" containerID="30e6fe3afab65597caa31dc9566ca49b39d65338a78357367c6cb3c4003713ee" Oct 02 12:01:21 crc kubenswrapper[4771]: I1002 12:01:21.683007 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:01:21 crc kubenswrapper[4771]: E1002 12:01:21.683958 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:01:36 crc kubenswrapper[4771]: I1002 12:01:36.683411 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:01:36 crc kubenswrapper[4771]: E1002 12:01:36.684838 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:01:49 crc kubenswrapper[4771]: I1002 12:01:49.684426 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:01:49 crc kubenswrapper[4771]: E1002 12:01:49.685275 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:02:02 crc kubenswrapper[4771]: I1002 12:02:02.681767 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:02:02 crc kubenswrapper[4771]: E1002 12:02:02.682775 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:02:14 crc kubenswrapper[4771]: I1002 12:02:14.681891 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:02:14 crc kubenswrapper[4771]: E1002 12:02:14.683413 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:02:27 crc kubenswrapper[4771]: I1002 12:02:27.688360 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:02:27 crc kubenswrapper[4771]: E1002 12:02:27.690153 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:02:39 crc kubenswrapper[4771]: I1002 12:02:39.687816 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:02:39 crc kubenswrapper[4771]: E1002 12:02:39.688872 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:02:51 crc kubenswrapper[4771]: I1002 12:02:51.682355 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:02:51 crc kubenswrapper[4771]: E1002 12:02:51.683651 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:03:05 crc kubenswrapper[4771]: I1002 12:03:05.682747 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:03:05 crc kubenswrapper[4771]: E1002 12:03:05.684343 4771 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m4mpt_openshift-machine-config-operator(52d686b1-d29a-42c4-97ce-e239a0c680b0)\"" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" podUID="52d686b1-d29a-42c4-97ce-e239a0c680b0" Oct 02 12:03:20 crc kubenswrapper[4771]: I1002 12:03:20.684400 4771 scope.go:117] "RemoveContainer" containerID="1aecf491329f58d346566f6cecf308cb898034923d41163272fe95c97e979096" Oct 02 12:03:21 crc kubenswrapper[4771]: I1002 12:03:21.965613 4771 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4mpt" event={"ID":"52d686b1-d29a-42c4-97ce-e239a0c680b0","Type":"ContainerStarted","Data":"e91c37f92e125c6c5263159ec23627c05d27218a3cab38712a56c2d7662cf414"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067465024024456 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067465025017374 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067443324016515 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067443324015465 5ustar corecore